Skip to main content
Advertisement
Browse Subject Areas
?

Click through the PLOS taxonomy to find articles in your field.

For more information about PLOS Subject Areas, click here.

  • Loading metrics

The neural correlates of context driven changes in the emotional response: An fMRI study

  • Brigitte Biró,

    Roles Conceptualization, Data curation, Formal analysis, Investigation, Methodology, Visualization, Writing – original draft, Writing – review & editing

    Affiliations NAP3.0-SE Neuropsychopharmacology Research Group, Hungarian Brain Research Program, Semmelweis University, Budapest, Hungary, Doctoral School of Psychology, ELTE Eötvös Loránd University, Budapest, Hungary, Institute of Psychology, ELTE Eötvös Loránd University, Budapest, Hungary

  • Renáta Cserjési,

    Roles Conceptualization, Methodology, Writing – review & editing

    Affiliation Institute of Psychology, ELTE Eötvös Loránd University, Budapest, Hungary

  • Natália Kocsel,

    Roles Investigation, Methodology, Project administration, Writing – review & editing

    Affiliation Institute of Psychology, ELTE Eötvös Loránd University, Budapest, Hungary

  • Attila Galambos,

    Roles Investigation, Methodology, Project administration, Writing – review & editing

    Affiliations Doctoral School of Psychology, ELTE Eötvös Loránd University, Budapest, Hungary, Institute of Psychology, ELTE Eötvös Loránd University, Budapest, Hungary

  • Kinga Gecse,

    Roles Investigation, Methodology, Writing – review & editing

    Affiliations NAP3.0-SE Neuropsychopharmacology Research Group, Hungarian Brain Research Program, Semmelweis University, Budapest, Hungary, Faculty of Pharmacy, Department of Pharmacodynamics, Semmelweis University, Budapest, Hungary

  • Lilla Nóra Kovács,

    Roles Investigation, Project administration, Writing – review & editing

    Affiliations Doctoral School of Psychology, ELTE Eötvös Loránd University, Budapest, Hungary, Institute of Psychology, ELTE Eötvös Loránd University, Budapest, Hungary

  • Dániel Baksa,

    Roles Investigation, Methodology, Writing – review & editing

    Affiliations NAP3.0-SE Neuropsychopharmacology Research Group, Hungarian Brain Research Program, Semmelweis University, Budapest, Hungary, Faculty of Pharmacy, Department of Pharmacodynamics, Semmelweis University, Budapest, Hungary

  • Gabriella Juhász,

    Roles Conceptualization, Funding acquisition, Methodology, Writing – review & editing

    Affiliations NAP3.0-SE Neuropsychopharmacology Research Group, Hungarian Brain Research Program, Semmelweis University, Budapest, Hungary, Faculty of Pharmacy, Department of Pharmacodynamics, Semmelweis University, Budapest, Hungary

  • Gyöngyi Kökönyei

    Roles Conceptualization, Data curation, Formal analysis, Methodology, Visualization, Writing – original draft, Writing – review & editing

    kokonyei.gyongyi@pharma.semmelweis-univ.hu, kokonyei.gyongyi@ppk.elte.hu

    Affiliations NAP3.0-SE Neuropsychopharmacology Research Group, Hungarian Brain Research Program, Semmelweis University, Budapest, Hungary, Institute of Psychology, ELTE Eötvös Loránd University, Budapest, Hungary, Faculty of Pharmacy, Department of Pharmacodynamics, Semmelweis University, Budapest, Hungary

Abstract

Emotional flexibility reflects the ability to adjust the emotional response to the changing environmental context. To understand how context can trigger a change in emotional response, i.e., how it can upregulate the initial emotional response or trigger a shift in the valence of emotional response, we used a task consisting of picture pairs during functional magnetic resonance imaging sessions. In each pair, the first picture was a smaller detail (a decontextualized photograph depicting emotions using primarily facial and postural expressions) from the second (contextualized) picture, and the neural response to a decontextualized picture was compared with the same picture in a context. Thirty-one healthy participants (18 females; mean age: 24.44 ± 3.4) were involved in the study. In general, context (vs. pictures without context) increased activation in areas involved in facial emotional processing (e.g., middle temporal gyrus, fusiform gyrus, and temporal pole) and affective mentalizing (e.g., precuneus, temporoparietal junction). After excluding the general effect of context by using an exclusive mask with activation to context vs. no-context, the automatic shift from positive to negative valence induced by the context was associated with increased activation in the thalamus, caudate, medial frontal gyrus and lateral orbitofrontal cortex. When the meaning changed from negative to positive, it resulted in a less widespread activation pattern, mainly in the precuneus, middle temporal gyrus, and occipital lobe. Providing context cues to facial information recruited brain areas that induced changes in the emotional responses and interpretation of the emotional situations automatically to support emotional flexibility.

Introduction

Emotional flexibility refers to the ability to modulate one’s emotional responses to fit the changing demands of the environmental context, and, thus, to change–generate, inhibit, down- or upregulate–one’s initial emotional responses according to the contextual demands [1,2]. Context can automatically direct emotional processing and can easily override facial expressions [3]. Learned associations between emotional responses and contexts lead to the appraisal of situations and shape the emotional responses and/or the regulatory processes [4,5]. For illustration, crying at a funeral represents sadness, in contrast with crying when achieving great success it represents happiness or pride. Thus, the meaning of a stimulus may depend on the context and may change when the context changes, so one of the key elements of emotional flexibility is shifting between meanings to adapt our behavior, e.g., our emotional response to the context. In short, to give an appropriate emotional response after context modifications [6,7].

Decoding emotions from faces has been extensively studied, but there is now a large body of evidence proving that the situational context (e.g., physical and social environment), along with the emotional/social knowledge of the perceiver about the situation will automatically guide the perception [8], causing even radical categorical changes in the perceived emotion (e.g., pride instead of sadness) [9]. Contextual information is processed and integrated with facial affective information in the early phase of perception [10]; thus, even the perception of a basic facial emotion can be categorically changed automatically by the context [9].

There are many ways to test the effect of context on emotion perception experimentally. For instance, knowledge about the situation of the observed person can be manipulated by semantic-linguistic labels [3,11] and information given before presenting even neutral faces [12]. Emotional faces can also be presented on different, even artificial backgrounds [13,14], or in naturalistic scenes. Results of electrophysiological [10] and magneto-encephalographic [15] studies suggest that facial perception is influenced by contextual cues even in the early stage of visual processing. On the basis of the available evidence, Aviezer and colleagues [16] conclude that context does not simply have a modulating effect on the processing and perception of emotions, but can actually lead to a categorical shift in the perception of an emotional expression.

Research on cognitive reappraisal can also help to understand the impact of context on emotional information processing. In reappraisal studies [17,18], when participants are asked to give a different meaning to a negative or positive stimulus, they are instructed to create a new cognitive context for the stimulus. Shifting from emotional to non-emotional or from negative to positive meaning (or vice versa) definitely alters the emotional trajectory, causing changes in the initial emotional responses [19]. In a multi-level framework, proposed by Braunstein, Gross, and Ochsner [20] reappraisal studies are considered to address controlled emotion regulation with explicit regulatory goals: participants are instructed to regulate their emotions (explicit regulatory goal) using effortful processes to change the cognitive context (controlled processes). However, reappraisal or shift in meaning, or more generally, emotion regulation can happen without explicit regulatory goals and/or in a more automatic manner (see [20,21]).

To extend previous works, our aim was to investigate a shift or change in emotional perception triggered by the context that occurs automatically and without explicit instructions to change. We used the Emotional Shifting Task (EST) [22,23], in which pairs of pictures are presented: the first one is a small detail of the second picture, as depicted in Fig 1. The presentation of the first picture, which is a decontextualized part of the whole picture, generates an emotional response, but when this picture is put into a context, the context itself may cause a change (and in some cases a shift) in the meaning and valence of the stimulus. For instance, a picture of a smiling girl is generally evaluated as a positive stimulus but if it turns out that she is smiling while bullying a peer, it will probably cause a shift in the evaluation toward a negative direction automatically and without any explicit regulatory goals. This technique was developed first by Munn in 1940 [24], who selected 16 pictures depicting emotional expressions from Life and Look magazines and then prepared two sets of pictures: one set with the full picture including the context as well, another set with only the face from the full picture. Munn found that the extra information of the context could change the judgment of an emotional face when participants were asked to name the emotion appearing on the face. The EST also uses naturalistic scenes as well in order to mimic real-life emotion perception and increase the ecological validity of our task.

thumbnail
Fig 1. Schematic representation of a sequence of trial in the Emotional Shifting Task.

An example for shifting from positive to negative emotion.

https://doi.org/10.1371/journal.pone.0279823.g001

Our aim was to explore the neuronal responses to the previously seen decontextualized photograph (depicting emotions using primarily facial and postural expressions) when it was presented in a naturalistic context. We expected that areas involved in processing of facial and contextual cues were going to be recruited by our task. For instance, studies using facial expressions in investigating emotion perception found that facial expressions activated the so-called face-selective regions, including the inferior occipital gyrus (occipital face area, OFA), lateral fusiform gyrus (fusiform face area), and posterior superior temporal sulcus (pSTS) as core regions of a widely distributed network [8]. Other brain regions such as the amygdala, anterior inferior temporal cortex [25], insula, and inferior frontal gyrus [26] also play a role in facial emotional processing [27]. As for the context, it is well established that complex social situations can easily trigger mentalizing, i.e., inferring mental and affective states to others (often called theory of mind) [28,29] and/or empathic responding, i.e., vicariously experiencing the feelings and emotional states of other people [30]. Core brain regions of the neural network of mentalization are the bilateral temporoparietal junction (TPJ) and medial prefrontal cortex (mPFC) [31] whereas the core network of empathy involves the dorsal anterior cingulate cortex, anterior midcingulate cortex and supplementary motor area (SMA) [32].

The EST task we used allowed us to distinguish different conditions as presented in Table 1. On the basis of two dimensions, which are (1) the valence (positive or negative) of the initial emotional stimulus and (2) the valence (positive or negative) of its contextualized presentation, four different types of automatic changes can be targeted in emotional response. We argue that emotional flexibility can be investigated with this design as Coifman and Summers [2] pointed out the initial emotional response (to the decontextualized pictures) must be modulated (upregulate the emotions or shift) to fit to the context.

thumbnail
Table 1. Understanding emotional flexibility in terms of context-driven emotional response.

https://doi.org/10.1371/journal.pone.0279823.t001

By shift we refer to categorical changes where the valence of the initial emotional response is reversed, i.e., the initial negative emotional response becomes positive and vice versa. The term upregulation is used when the valence of the same initial emotional response is increased by the context; thus, a negative stimulus becomes more negative, or a positive one becomes more positive. Accordingly, the EST contained two shift and two non-shift (upregulation) conditions.

On the basis of the theory by Saxe and Houlihan [33] different emotional responses could be expected to stimuli in context vs. without context. They argue that forward inferences are used to attribute emotions to the target when an emotional expression is processed in a context; thus, we automatically infer that the cause of the emotional state of the target reflected in their emotional expressive behavior is the context/event. On the basis of this, we expected that context itself would recruit areas involved in emotional processing and understanding complex social situations; thus, first we simply compared the neural responses to whole pictures vs. decontextualized (cropped) pictures. We refer to this as a general context effect in our study. Then we used this activation map as an exclusive mask to be able to explore the four different types of automatic changes specifically in emotional responses. It allowed us to explore neural activation to changes in the meaning triggered by the context as a passive cue independent of the context vs. no context differences. On the basis of previous studies, we hypothesized that prefrontal regions, especially the mPFC and dorsolateral prefrontal cortex (dlPFC) [17,34] were going to be recruited when the context induced a shift in the emotional valence of the pictures. More specifically, on the basis of a recent study on the automatic regulation of negative emotions by Yang and coworkers [35] we expected that visual areas, striatal areas, precentral/postcentral gyri and dlPFC would be activated when context resulted in a shift from negative towards positive meaning.

Method

Participants

Thirty-two healthy adult volunteers recruited through social media sites and journal advertisements were included in the present study; however, one participant was excluded from the first level analysis due to excessive movement during the fMRI measure; thus, the final data of 31 participants, 18 females and 13 males (mean age: 24.44 ± 3.4), were analyzed. The participants were right-handed, as assessed by the Edinburgh Handedness Inventory [36], and had normal or corrected-to-normal vision. All participants were examined by a senior psychiatrist and neurologist and were excluded with any history of psychiatric or neurological disorders or chronic medical conditions.

The present study was approved by the Scientific and Research Ethics Committee of the Medical Research Council (Hungary), and written informed consent was received from all subjects in accordance with the Declaration of Helsinki.

Psychological task: The Emotional Shifting Task

The EST [22] consists of 24 picture pairs. In each pair, the first picture is always a smaller detail from the second (whole) picture. In most cases the cropped image expressed emotion primarily through facial expression and/or posture. The valence of the firstly presented picture either remains or changes when it is placed into a context, and so should change the elicited emotion (Fig 1). For the upregulation conditions (P1P2 and N1N2), pictures were selected from the International Affective Picture System [37]. Their identification numbers were 1340, 2091, 2141, 2205, 2216, 2340, 2530, 2700, 6242, 6838, 8497, and 9050. For the shift conditions (P1N2 and N1P2) pictures were selected from the internet. Six criteria were used to select the images: (1) free for non-commercial use, (2) depicting social interactions, (3) evoking an emotional response without being shocking or extreme, (4) not depicting famous person(s), (5) eligible for shifting conditions, i.e., the valence of facial expression and the whole picture should be opposite, and (6) the images should represent as many different situations as possible.

After each pair, a happy and a sad smiley/emoji appeared on the screen (Fig 2), and participants had to choose one of them by pressing the corresponding button to indicate the valence (positive or negative) of the second (whole) picture. We decided to use emojis in the scanner to mimic the two endpoints of valence ratings in Self-Assessment Manikin [37].

Four conditions were defined in the task: two conditions, in which participants were required to alter their emotions either from positive to negative (P1N2) or from negative to positive (N1P2), and two conditions, where no shift (but upregulation) was expected in the valence (i.e. both pictures presented were either positive (P1P2) or negative (N1N2)). Each condition consisted of six pairs of pictures, presented in pseudo-random order. During functional magnetic imaging, two behavioral variables were registered: the reaction times (RT) of the selection of the emoji and the number of “correct” answers. An answer was considered correct if the valence of the secondly presented picture matched the valence of the selected smiley. Stimulus presentations and data registrations were conducted in E-Prime 2.0 Software (Psychology Software Tools, Inc., Pittsburgh, PA, USA).

Procedure

Data collection included three steps. First, participants completed a short practice session, which was explained and presented on a laptop outside of the scanner and consisted of three pairs of stimuli that were not used in the task and included shift and non-shift conditions as well. In this part, participants could read the instructions and ask their questions in case of uncertainty regarding the instructions or the operation of the task.

In the next step, participants were instructed to get as emotionally involved in the presented situations as possible while viewing the pictures in the scanner. To measure a baseline brain activity, a white fixation cross was presented on a black background at the beginning and at the end of the task for 20 seconds. Each emotional stimulus was shown for 8 seconds. The timing was based on laboratory pilot studies and previous studies [11]. To avoid artifacts due to expectations, fixation crosses were presented with altered timing (from 5 to 11 seconds, mean presentation time: 8 seconds) before each emotional stimulus. At the end of each trial, the answer screen was presented for 4 seconds (Fig 2).

Lastly, a post-test was filled out after the fMRI measure (outside the scanner) to examine whether participants observed changes in the valence of the pictures. In this part, the pairs of pictures were presented in the same order on a laptop, and participants were asked to rate them on a 7-point Likert scale. Valence and arousal were measured from 1 to 7 (1 being very unpleasant and 7 very pleasant; 1 being calm and 7 very excited, respectively).

FMRI acquisition

The functional MRI data collection was carried out by a SIEMENS MAGNETOM Prisma syngo MR D13D 20 channels headcoil 3T scanner. A BOLD-sensitive T2*-weighted echo-planar imaging sequence was used (TR = 2220ms, TE = 30 ms, FOV: 222) with 3 mm × 3 mm in-plane resolution and contiguous 3-mm slices providing whole-brain coverage. Four hundred and nine volumes were acquired during the task. For the structural data a series of high-resolution anatomical images were acquired before the functional imaging using a T1-weighted 3D TFE sequence with 1 × 1 × 1 mm resolution.

Statistical analysis of self-report and post-test data

To analyze demographic and behavioral data SPSS version 28.0 (IBM SPSS, IBM Corp, Armonk, NY, USA) was used, and descriptive and non-parametric statistics were performed. As the distribution of valence and arousal ratings was non-normal, we used Wilcoxon Signed Rank Test to compare the valence and arousal ratings of the first and second pictures in each condition (P1N2, N1P2, P1P2, and N1N2). However, as it was easier to interpret changes in means than in ranks, we repeated these analyses using a series of bootstrapped paired t-tests. A repeated measures ANOVA was performed on the reaction times collected during the fMRI scan.

FMRI data analyses

Preprocessing.

Statistical Parametrical Mapping (SPM12) analysis software package (Wellcome Department of Imaging Neuroscience, Institute of Neurology, London, UK; http://www.fil.ion.ucl.ac.uk/spm12/ implemented in Matlab 2016b (Math Works, Natick, MA, USA) was used to analyze the imaging data. Preprocessing contained the following steps: realignment, co-registration to the structural image, segmentation, normalization in the Montreal Neurological Institute (MNI) space, and spatial smoothing with an 8-mm full-width half-maximum Gaussian kernel. These steps of preprocessing were performed on the functional images. Finally, a visual inspection of the pictures took place to exclude the poor-quality images.

First level model.

During first-level analyses, BOLD (blood oxygenation level-dependent) hemodynamic responses were modeled in a general linear model. In the event-related single subject analysis fixation screens, both stimuli (positive and negative), the disposition of the shift (the valence of the first stimuli: positive/negative, and the nature of the condition: shift/non-shift condition) and the two possible answers (happy and sad emojis) were modeled as separate regressors of interest. High-pass temporal filtering with a cut-off of 128 s was included in the model to remove the effects of low-frequency physiological noise, and serial correlations in data series were estimated using an autoregressive AR (1) model. Motion outliers (threshold of global signal > 3 SD and motion > 1 mm) were identified with the Artifact Detection Tools (ART; www.nitrc.org/projects/artifact_detect/), and the six motion parameters were used as regressors of no interest in the fMRI model.

Four contrasts were created to analyze whether an increased activation could be detected to stimuli that were placed into a context (2nd picture) compared to the ones without contextual background (1st picture), and also focusing on the valence of the stimuli (see Table 1).

Second-level analyses.

During second-level analyses (whole brain t-test) the threshold was set to p< .05 family-wise error (FWE) corrected for multiple comparisons. The automated anatomical labeling atlas (aal) [38] was used to anatomically identify the activated clusters, whereas the MNI 152 template brain provided in MRIcroGL was used to visualize statistical maps http://www.mccauslandcenter.sc.edu/mricrogl/ [39].

Results

Behavioral results

Descriptive statistics.

To track the changes in valence and arousal in picture pairs, ratings of the valence and arousal values registered in the post-task after scan were analyzed (S1 Table). Answers were compared to the post-task, and the valence ratings of all the pictures were in the expected directions, namely positive pictures (P1; P2) were rated more pleasant and negative pictures (N1; N2) more unpleasant. Results of the Wilcoxon Signed Rank Test (and bootstrapped paired t-test) showed the significant differences between the mean valence and arousal values within conditions according to which valence ratings significantly differed in each condition (P1N2; P1P2; N1P2; N1N2), similarly, arousal ratings showed a significant increase for the second pictures (P2, N2), compared to the first ones (P1, N1) in each condition (see S1 Table, Figs 3 and 4).

thumbnail
Fig 3. Changes in valence ratings in the post-task.

Note. P1: First picture of the picture pairs is positive. N1: First picture of the picture pairs is negative. P2: Second picture of the picture pairs is positive. N2: Second picture of the picture pairs is negative. * p < .001.

https://doi.org/10.1371/journal.pone.0279823.g003

thumbnail
Fig 4. Changes in arousal ratings in the post-task.

Note. P1: First picture of the picture pairs is positive. N1: First picture of the picture pairs is negative. P2: Second picture of the picture pairs is positive. N2: Second picture of the picture pairs is negative. * p < .001.

https://doi.org/10.1371/journal.pone.0279823.g004

Descriptive data, collected during the EST, provided information on the mean reaction time values given in the different contrasts (see S2 Table) and also on the “accuracy” of the answers of the participants (whether they picked the expected smiley/emoji showed on the screen). This accuracy was 95.54% (range: 79–100%). A repeated measures ANOVA was performed on the reaction times collected during the fMRI scan, resulting in a significant difference across the four contrasts (F3, 90 = 16.273, p< .001). Post hoc analyses showed that subjects pressed the button more slowly in trials of shifting from positive to negative (P1N2) compared to the other three types of trials. Reaction times in the trials when both pictures were negative (N1N2) were longer, compared to the trials when both pictures were positive (P1P2) and to the trials of shifting from negative to positive (N1P2) (see S2 Table and Fig 5).

thumbnail
Fig 5. Mean reaction times (and standard deviations in milliseconds) to the second picture of the picture pairs in the scanner by the type of picture pairs in the Emotional Shifting Task.

Note. P1: First picture of the picture pairs is positive. N1: First picture of the picture pairs is negative. P2: Second picture of the picture pairs is positive. N2: Second picture of the picture pairs is negative. * p < .05.

https://doi.org/10.1371/journal.pone.0279823.g005

Task-related activations

Main effect of context.

The main effect of the context was checked by comparing the increased brain activations of the second images (full pictures with the context) to the firstly presented images (pictures without context) regardless of valence changes. Widespread activations were found in the brainstem, lingual and fusiform gyri, precuneus, calcarine, middle and superior temporal gyri, middle and superior occipital gyri, inferior parietal gyrus, middle, superior, medial, inferior frontal gyri, precentral gyrus, SMA, anterior cingulate (ACC) and postcentral gyrus (S3 Table and S1 Fig).

From positive to negative: Activation to positive pictures in a negative context compared to positive pictures without context (P1-N2).

To reveal the increased brain activation specific to the change in the meaning from positive to negative triggered by context, the context main effect regions were used as an exclusive mask on the results of the widespread increased activations to positive pictures in a negative context compared to positive pictures without context (Table 2 and Fig 6). Thus, the regions activated outside the mask were the superior medial frontal gyrus, inferior orbitofrontal gyrus, superior temporal pole, middle and superior temporal gyrus, middle occipital gyrus, SMA, anterior cingulum, thalamus, caudate and amygdala.

thumbnail
Fig 6. Activated regions when the context categorically changed the valence of the emotional stimuli (from negative to positive and from positive to negative) after excluding the general effect of context.

Increased activations when a negative stimulus was put in a positive context are shown in green whereas increased activations when a positive stimulus was put in a negative context are shown in red. Coordinates are in the Montreal Neurological Institute (MNI) space. Statistical maps were visualized on the MNI 152 template brain provided in MRIcroGL [39].

https://doi.org/10.1371/journal.pone.0279823.g006

thumbnail
Table 2. Brain regions showing increased activation to positive pictures in a negative context compared to positive pictures without context and brain regions showing significantly increased activation to negative pictures in a positive context compared to negative pictures without context.

https://doi.org/10.1371/journal.pone.0279823.t002

From negative to positive: Brain activations to negative pictures in a positive context compared to negative pictures without context (N1P2).

In picture pairs, where the secondly presented positively valenced stimuli (P2) were compared to the firstly presented negatively valenced stimuli (N1), increased activations were found in the middle temporal gyrus, middle cingulum, middle occipital gyrus, precuneus and calcarine when we used the above mentioned exclusive mask (Table 2 and Fig 6).

From positive to positive: Brain activations when both the first and second pictures were positive (P1P2).

Increased BOLD signals were found in the visual areas including the calcarine and the lingual gyrus (Table 3) when positive pictures in a positive context were compared to positive pictures without context when using the exclusive mask.

thumbnail
Table 3. Brain regions showing significantly increased activation to positive pictures in a positive context were compared to positive pictures without context (P1P2 picture pairs), and those with significantly increased activation to negative pictures in a negative context were compared to negative pictures without context (N1N2 picture pairs).

https://doi.org/10.1371/journal.pone.0279823.t003

From negative to negative: Brain activations when both the first and second pictures were negative (N1N2).

The superior occipital gyrus showed an increased activation (Table 3) to negative pictures in a negative context compared to negative pictures without context, and the context main effect regions were used as an exclusive mask.

Discussion

In the present study we aimed to measure emotional flexibility, defined here as a change in the emotional response elicited by a specific context. This task is built on the notion that the context may give an entirely different interpretation to the stimulus, resulting in even categorical changes in the valence and/or arousal of the elicited emotion [2]. In our task, the context appeared as a passive cue that induced changes in the emotional responses and interpretation of the emotional situations automatically; thus, this task is considered to explore changes in the spontaneous emotional output guided by the context. Thus, on the basis of the review by Coifman and Summers [2], we argue that EST is an appropriate task to capture emotional flexibility. Post-task data on the valence, arousal and changes in BOLD responses support that this task can induce changes in the emotional response.

Context effect on valence, arousal rating and reaction times

To understand the outcome of the fMRI results better, valence and arousal ratings of the emotional stimuli were collected during the post-task (out of scanner post-task): participants were asked to rate the pictures used in the fMRI task after the scan. Significant changes in the valence and arousal ratings of the stimuli were observed after they were placed into a context, indicating that participants reinterpreted the emotional stimuli. These changes in the valence and arousal were detected in all four types of picture sets, indicating that the context itself did not only shape the categorization of the emotional states [9], but it also might affect valence and intensity (arousal). The context, or more precisely the appraisal of overall context (e.g. bullying, being in a hospital with a sick person, or childbirth), i.e., the semantic features, might provide extra affective information, as it gives an explanation for the emotion; thus, it guides our interpretation [33]. For instance, seeing a crying woman in hospital compared to seeing just a crying woman, could elicit a more intense emotional response, as it can activate additional emotional meaning or knowledge such as her relative being sick.

Reaction times in the scanner were also registered and showed that participants performed more slowly to the second picture in trials of shifting from positive to negative compared to the other three types of trials. Reaction times to the second picture were longer in the trials when both pictures were negative compared to trials when both pictures were positive or when shifting from negative to positive was required. Thus, reaction time was longer when the second picture was negative compared to when it was positive. We did not ask our participants to choose an emoji after the first picture in the scanner; thus, it limits our understanding.

This result is in line with the findings of Sakaki and coworkers [40] who presented negative, neutral and positive pictures in their study and found that participants had longer reaction times to negative pictures compared to neutral or positive pictures in their task that required semantic processing. On the basis of the results, Sakaki et al. [40] concluded that facing pictures of negative emotional events affected or interfered with the semantic processing of the following stimuli even more than perceptual processing.

General context effect on brain activation

Overall, we found increased occipital cortex activation when emotional stimuli were put into context, indicating heightened perception and attention [41] presumably evoked by the information about the social and physical environment surrounding the expressor, along with existing emotional knowledge about the context. The activation of the lateral occipital cortex has been found in studies investigating emotional scene processing [42], and the role of calcarine in visual information processing [43,44] and in visual-imagery processes is well-documented [45].

In our task context brought new knowledge (information) on the emotional state of the protagonist. Indeed, many of the activated areas such as the middle temporal gyrus (MTG), fusiform gyrus, SMA [34], temporal pole [46] caudate [47], brainstem [48] and thalamus [49,50] suggest that the social and emotional meaning of facial/postural information, possibly along with other faces on the picture, required an increased emotional reprocessing when the context appeared.

The mental states we attribute to others and the extent to which we resonate with their emotional states can guide our behavior in complex social situations. Accordingly, the contextual presentation of emotional stimuli may activate brain areas involved in mentalization and/or empathic response. Thorough investigation of activation maps revealed that the MTG activation extended to the TPJ. Several meta-analyses suggest [31,51,52] that TPJ is one of the core regions associated with social cognition, or more specifically mentalizing (often called theory of mind) [28,53], and TPJ is also activated in empathy studies [51]. Furthermore, the precuneus was also activated, this area was previously associated with affective mentalizing [54] i.e., inferring affective states of others, occasionally correspondingly called cognitive empathy. The recruitment of TPJ and the precuneus in our task might suggest that when context was added to the first picture the observers (our participants) reflected on the emotional meaning of the situation.

We detected increased anterior insula activation when context was added to the first picture. Its role in processing interoceptive information that is a key to representing emotional experience [55] is well-established, and insula activation correlated with self-reported arousal [56]. In a recent study [57] activations in other regions, such as in the superior temporal sulcus, fusiform gyrus and lateral occipital cortex, have also been associated with arousal. Note that our post-test after scanning revealed that the second (whole) pictures were more arousing compared to the first one.

We also found increased activation in the right inferior frontal gyrus and lateral orbitofrontal regions when context was added to the first picture. According to a meta-analysis [58] major overlaps in both regions along with the insula and temporal structures can be seen in emotional processing, interoceptive signaling, and social cognition, supporting previous neuroimaging data and our previous expectations that adding a context will recruit areas involved in emotional processing and understanding complex social situations. These results indirectly support the theory of Saxe and Houlihan [33] proposing that information about the event in which the target is expressing emotion is used as a cause for inferring target’s emotion(s).

Specific context effect

Placing positive stimulus in negative context.

After excluding the general effect of context by using a mask with activation to context vs. no-context, the automatic shift from positive to negative valence was associated with increased BOLD response in dorsomedial PFC, SMA, lateral orbitofrontal cortex (OFC), rectus, caudate, thalamus, amygdala, and MTG. Many of the activated areas such as the MTG, SMA [34], orbitofrontal cortex [46] caudate [47], and thalamus [49,50] suggest that the context required an increased emotional reprocessing (beyond the general context effect) when it made a previously positive picture negative.

In the field of emotion, activations in the lateral OFC in our study correspond to face selective part of OFC observed in a face discrimination reversal task when a formerly correct face was no longer the correct choice [59]. More specifically, in that study when a correct face was chosen, its expression turned into a smile, but when the wrong face was chosen, it turned into an angry face. In that study this part of OFC was activated when a formerly correct face was no longer the correct choice, so instead of a smile its choice resulted in an angry face. Thus, activation in this part of OFC was proposed to be error-related, as there was a discrepancy between the expected and perceived feedback. This area was recruited when face expression signaled a need for behavior change (i.e., change in the choices) [60]. On the basis of this, seeing a formerly positive face expression in an overall negative context–e.g., a smiling girl in a bullying context–also triggers error-related processes and might signal a need for behavior change. In our study, the biggest change in valence according to our post-test results emerged when positive pictures were put into a negative context. This change, or more precisely the shift from a positive to a negative meaning might require the reformulation of mental representations as well reflected in the increased activations in the supplementary motor area [34].

Medial PFC, especially its dorsomedial part has been activated in studies investigating empathy with other regions such as the SMA and thalamus [51]. In addition, medial prefrontal regions are recruited in reappraisal studies [61]. For instance, previous studies suggest that the regulation of negative emotions needs the allocation of cognitive resources provided by the dorsomedial prefrontal and dorsal cingulate gyrus [62]. The automatic shift from positive to negative in the meaning of the stimulus in our study might be accompanied by the recruitment of regions providing cognitive sources to regulate the resulting negative emotions. However, we did not find activation in the dorsolateral prefrontal cortex commonly observed in reappraisal studies [17,34,63].

We detected small activation in the amygdala that is often associated with negative/fearful emotional experiences, faces [64], emotional events and personal affective importance [65] or motivational relevance [66]. As this was the first fMRI study using the EST, we decided to present all significant activations. However, the cluster size was too small to interpret this result in the context of shifting.

The above results suggest that after excluding the general effect of context, shift in the meaning of a positive stimulus to negative induced by the context was supported in our task by areas involved in emotional processing, reformulation of mental representations, mentalizing, empathy and, to some extent, error-related process and cognitive control.

Placing negative stimuli in positive context.

Interestingly, trials where a negative facial expression was placed into a positive context resulted in a less widespread activation pattern compared to trials where a positive facial expression turned out to have a negative meaning. The increased activation of MTG and fusiform gyrus suggested that this shift was also associated with increased emotional processing [67], whereas the recruitment of the precuneus also supports that affective and cognitive processing of affective mental states was also increased in the participants [51]. However, in this case, we did not detect activations in the medial or lateral part of the frontal cortex, suggesting that this automatic shift, or the resulting positive emotions, did not require extra cognitive resources or regulation at least in our task. Thus, the differences and similarities between effortful and automatic regulation of negative emotions require further studies.

Our results differ from those of Yang and colleagues [35]. Although they investigated spontaneous recovery from a negative emotional state as an implicit form of emotion regulation, we used context as a passive cue to trigger a change in emotional response, suggesting that different forms of automatic emotion regulation need to be tested in further studies.

Placing negative stimuli in negative context and positive stimuli in positive context.

Interestingly, when a positive stimulus was placed into a positive context, or a negative stimulus was placed into a negative context, increased activation primarily in the occipital regions (calcarine, lingual gyrus and superior occipital lobule) was detected. In the literature, upregulating negative [68] or positive emotions [69] are mainly examined in reappraisal studies where participants are instructed to use certain tactics; thus, effortful and controlled processes are targeted. In our task, upregulation of emotions–confirmed by the valence and arousal ratings–was induced by the context that did not require effortful processes, which might explain our results. However, only small clusters were found in these two conditions after masking, so we should interpret these results with caution.

Limitations

We asked participants to rate the valence and arousal of pictures during the post-task, not during the fMRI measure; however, results showed that the emotional valence ratings of the second, whole pictures changed in the expected direction. In addition, we did not record the eye movements of the participants, so we cannot rule out the possibility that the four different conditions differed in the amount of eye movements, and that this may have affected our results.

In order to better understand the impact of the complex context in an emotional situation, adding conditions that include the context without the face might have been useful [70]. That would have allowed us to see if adding the face to the context would modify the valence beyond the information already evident in the context itself. However, the aim of the study was not to study facial and contextual information processing per se, but to use the context as a passive cue that promotes a shift in the meaning of certain set of images, therefore supporting emotion regulation.

Participants did not have to choose from emojis after the first picture, so we could not calculate reaction time differences within the pairs, but simply compared reaction times to second pictures, which limits our understanding; however, reaction time differences were not in the focus of our study. Additionally, it would have been ideal to put the same first picture in a negative and a positive context as well to directly compare how different contexts might influence the emotional processing of the same emotional stimulus.

Another limitation could be that participants might have had a certain expectation regarding the context that might have appeared in the activations to the second pictures. However, to avoid this or to decrease its possibility, participants were specifically instructed to solely focus on the stimuli on hand.

For the non-shift trials, stimuli were selected from the IAPS [37] database, whereas pictures for the shift trials were collected from the internet, so they are not from a standardized set of emotional stimuli; however, they went through several pilot studies [22]. Individual differences in the emotional reactivity, empathy, or ToM might affect the perception of emotional stimuli [71], but we did not assess these characteristics of our participants.

Conclusion

We aimed to capture emotional flexibility by a task using context as a cue to trigger (an automatic) change in emotional response. The affective information and the social knowledge activated by the context had a major impact on the neural processing of the emotional visual stimuli. Presenting previously seen decontextualized emotional stimulus in a context recruited areas involved in emotional processing and understanding complex social situations, probably indicating that the context itself narrows the probability of emotions previously attributed to the decontextualized stimulus [72]. Thus, information about the context might be used as a cause of emotions [33].

Additionally, our results highlight that sensitivity and appropriate responses to context depend on many different processes; thus, emotional inflexibility may stem from different underlying mechanisms. Therefore, understanding emotional inflexibility in psychopathologies requires the dissection of these underlying mechanisms first.

Supporting information

S1 Fig. General context effect: Full pictures with the context vs. firstly presented images (pictures without the context) at p< .05, family-wise error (FWE) corrected for multiple comparison.

Coordinates are in Montreal Neurological Institute (MNI) space. Statistical maps were visualized on the MNI 152 template brain provided in MRIcroGL [39].

https://doi.org/10.1371/journal.pone.0279823.s001

(DOCX)

S1 Table. Valence and arousal values for the pictures of the post-task by contrasts.

P1: First picture of the picture pairs is positive. N1: First picture of the picture pairs is negative. P2: Second picture of the picture pairs is positive. N2: Second picture of the picture pairs is negative. + Results of the bootstrapped paired t-test. ++ Results of the Wilcoxon Signed Rank Test. *p < .001. Valence and arousal were measured from 1 to 7 (1 being very unpleasant and 7 very pleasant; 1 being calm and 7 very excited, respectively).

https://doi.org/10.1371/journal.pone.0279823.s002

(DOCX)

S2 Table. Mean reaction times (in milliseconds) to the second picture in the scanner by the type of picture pairs in the Emotional Shifting Task.

P1: First picture of the picture pairs is positive. N1: First picture of the picture pairs is negative. P2: Second picture of the picture pairs is positive. N2: Second picture of the picture pairs is negative. Different letters (a, b, c) represent significant (p < .05) difference between mean scores, whereas the same letters represent non-significant difference between mean scores according to the paired post hoc test of repeated measure of ANOVA.

https://doi.org/10.1371/journal.pone.0279823.s003

(DOCX)

S3 Table. General context effect: Increased activations to the 2nd pictures with context compared to the 1st pictures without context.

L = left; the initial statistical threshold was set to p< .05, family-wise error (FWE) corrected for multiple comparison.

https://doi.org/10.1371/journal.pone.0279823.s004

(DOCX)

Acknowledgments

The authors thank Mária Kelner for the drawing the first figure and Tamás Smahajcsik-Szabó for the figures with behavioural data.

References

  1. 1. Beshai S, Prentice JL, Huang V. Building blocks of emotional flexibility: Trait mindfulness and self-compassion are associated with positive and negative mood shifts. Mindfulness. 2018;9(3):939–48.
  2. 2. Coifman KG, Summers CB. Understanding Emotion Inflexibility in Risk for Affective Disease: Integrating Current Research and Finding a Path Forward. Frontiers in Psychology. 2019;10. WOS:000459769500001. pmid:30873087
  3. 3. Kayyal M, Widen S, Russell JA. Context Is More Powerful Than We Think: Contextual Cues Override Facial Cues Even for Valence. Emotion. 2015;15(3):287–91. WOS:000354544600004. pmid:25706831
  4. 4. Malooly AM, Genet JJ, Siemer M. Individual differences in reappraisal effectiveness: the role of affective flexibility. Emotion. 2013;13(2):302. pmid:23163706
  5. 5. Robinson MD, Watkins ER, Harmon-Jones E. Cognition and emotion: An introduction. In Robinson M. D., Watkins E., & Harmon-Jones E. (Eds.), Handbook of cognition and emotion (pp. 3–16). The Guilford Press; 2013.
  6. 6. Fu F, Chow A, Li J, Cong Z. Emotional flexibility: Development and application of a scale in adolescent earthquake survivors. Psychological trauma: theory, research, practice, and policy. 2018;10(2):246. pmid:28557483
  7. 7. Waugh CE, Thompson RJ, Gotlib IH. Flexible emotional responsiveness in trait resilience. Emotion. 2011;11(5):1059. pmid:21707168
  8. 8. Wieser MJ, Brosch T. Faces in context: a review and systematization of contextual influences on affective face processing. Frontiers in Psychology. 2012;3:471. pmid:23130011
  9. 9. Aviezer H, Hassin RR, Ryan J, Grady C, Susskind J, Anderson A, et al. Angry, disgusted, or afraid? Studies on the malleability of emotion perception. Psychological science. 2008;19(7):724–32. pmid:18727789
  10. 10. Hietanen JK, Astikainen P. N170 response to facial expressions is modulated by the affective congruency between the emotional expression and preceding affective picture. Biological Psychology. 2013;92(2):114–24. pmid:23131616
  11. 11. Deak A, Bodrogi B, Biro B, Perlaki G, Orsi G, Bereczkei T. Machiavellian emotion regulation in a cognitive reappraisal task: An fMRI study. Cognitive, Affective, & Behavioral Neuroscience. 2017;17(3):528–41.
  12. 12. Schwarz JM, Smith SH, Bilbo SD. FACS analysis of neuronal–glial interactions in the nucleus accumbens following morphine administration. Psychopharmacology. 2013;230(4):525–35. pmid:23793269
  13. 13. Righart R, De Gelder B. Rapid influence of emotional scenes on encoding of facial expressions: an ERP study. Social Cognitive and Affective Neuroscience. 2008;3(3):270–8. pmid:19015119
  14. 14. Righart R, De Gelder B. Recognition of facial expressions is influenced by emotional scene gist. Cognitive, Affective, & Behavioral Neuroscience. 2008;8(3):264–72. pmid:18814463
  15. 15. Morel S, Beaucousin V, Perrin M, George N. Very early modulation of brain responses to neutral faces by a single prior association with an emotional context: evidence from MEG. Neuroimage. 2012;61(4):1461–70. pmid:22525875
  16. 16. Aviezer H, Ensenberg N, Hassin RR. The inherently contextualized nature of facial emotion perception. Current Opinion in Psychology. 2017;17:47–54. WOS:000414464100009. pmid:28950972
  17. 17. Ochsner KN, Silvers JA, Buhle JT. Functional imaging studies of emotion regulation: a synthetic review and evolving model of the cognitive control of emotion. Annals of the New York Academy of Sciences. 2012;1251:E1. pmid:23025352
  18. 18. Ochsner KN, Gross JJ. Cognitive emotion regulation: Insights from social cognitive and affective neuroscience. Current Directions in Psychological Science. 2008;17(2):153–8. pmid:25425765
  19. 19. Gross JJ. The emerging field of emotion regulation: An integrative review. Review of General Psychology. 1998;2(3):271–99.
  20. 20. Braunstein LM, Gross JJ, Ochsner KN. Explicit and implicit emotion regulation: a multi-level framework. Social cognitive and affective neuroscience. 2017;12(10):1545–57. pmid:28981910
  21. 21. Gyurak A, Gross JJ, Etkin A. Explicit and implicit emotion regulation: a dual-process framework. Cognition and Emotion. 2011;25(3):400–12. pmid:21432682
  22. 22. Biro B, Kokonyei G, De Oliveira Negrao R, Dancsik A, Karsai S, Logemann HNA, et al. Interaction between emotional context-guided shifting and cognitive shifting: Introduction of a novel task. Neuropsychopharmacol Hung. 2021;23(3):319–30. pmid:34751083.
  23. 23. Lacroix A, Dutheil F, Logemann A, Cserjesi R, Peyrin C, Biro B, et al. Flexibility in autism during unpredictable shifts of socio-emotional stimuli: Investigation of group and sex differences. Autism. WOS:000736641700001. pmid:34957880
  24. 24. Munn NL. The effect of knowledge of the situation upon judgment of emotion from facial expressions. The Journal of Abnormal and Social Psychology. 1940;35(3):324–38.
  25. 25. Zhang H, Japee S, Nolan R, Chu C, Liu N, Ungerleider LG. Face-selective regions differ in their ability to classify facial expressions. Neuroimage. 2016;130:77–90. WOS:000372745600007. pmid:26826513
  26. 26. Haxby JV, Hoffman EA, Gobbini MI. The distributed human neural system for face perception. Trends in Cognitive Sciences. 2000;4(6):223–33. pmid:10827445
  27. 27. Duchaine B, Yovel G. A revised neural framework for face processing. Annual Review of Vision Science. 2015;1:393–416. pmid:28532371
  28. 28. Premack D, Woodruff G. Does the chimpanzee have a theory of mind? Behavioral and Brain Sciences. 1978;1(4):515–26.
  29. 29. Frith CD, Frith U. The neural basis of mentalizing. Neuron. 2006;50(4):531–4. pmid:16701204
  30. 30. Davis MH. Measuring individual differences in empathy: Evidence for a multidimensional approach. Journal of Personality and Social Psychology. 1983;44(1):113.
  31. 31. Molenberghs P, Johnson H, Henry JD, Mattingley JB. Understanding the minds of others: A neuroimaging meta-analysis. Neuroscience & Biobehavioral Reviews. 2016;65:276–91.
  32. 32. Fan Y, Duncan NW, de Greck M, Northoff G. Is there a core neural network in empathy? An fMRI based quantitative meta-analysis. Neuroscience & Biobehavioral Reviews. 2011;35(3):903–11.
  33. 33. Saxe R, Houlihan SD. Formalizing emotion concepts within a Bayesian model of theory of mind. Current Opinion in Psychology. 2017;17:15–21. WOS:000414464100004. pmid:28950962
  34. 34. Kohn N, Eickhoff SB, Scheller M, Laird AR, Fox PT, Habel U. Neural network of cognitive emotion regulation—an ALE meta-analysis and MACM analysis. Neuroimage. 2014;87:345–55. pmid:24220041
  35. 35. Yang Y, Zhang X, Peng Y, Bai J, Lei X. A dynamic causal model on self-regulation of aversive emotion. Brain Informatics. 2020;7(1):20. pmid:33296052
  36. 36. Oldfield RC. The assessment and analysis of handedness: the Edinburgh inventory. Neuropsychologia. 1971;9(1):97–113. pmid:5146491
  37. 37. Lang PJ, Bradley MM, Cuthbert BN. International Affective Picture System (IAPS): Technical manual and affective ratings. Washington, DC: NIMH Center for the Study of Emotions and Attention; 1997.
  38. 38. Tzourio-Mazoyer N, Landeau B, Papathanassiou D, Crivello F, Etard O, Delcroix N, et al. Automated anatomical labeling of activations in SPM using a macroscopic anatomical parcellation of the MNI MRI single-subject brain. Neuroimage. 2002;15(1):273–89. pmid:11771995
  39. 39. Rorden C, Brett M. Stereotaxic display of brain lesions. Behav Neurol. 2000;12(4):191–200. Epub 2001/09/25. pmid:11568431.
  40. 40. Sakaki M, Gorlick MA, Mather M. Differential Interference Effects of Negative Emotional States on Subsequent Semantic and Perceptual Processing. Emotion. 2011;11(6):1263–78. WOS:000297921200001. pmid:22142207
  41. 41. Keil A, Costa V, Smith JC, Sabatinelli D, McGinnis EM, Bradley MM, et al. Tagging cortical networks in emotion: A topographical analysis. Human Brain Mapping. 2012;33(12):2920–31. WOS:000310798800014. pmid:21954087
  42. 42. Sabatinelli D, Fortune EE, Li Q, Siddiqui A, Krafft C, Oliver WT, et al. Emotional perception: Meta-analyses of face and natural scene processing. NeuroImage. 2011;54(3):2524–33. pmid:20951215
  43. 43. Engel SA, Glover GH, Wandell BA. Retinotopic organization in human visual cortex and the spatial precision of functional MRI. Cerebral Cortex. 1997;7(2):181–92. pmid:9087826
  44. 44. Onitsuka T, Shenton ME, Salisbury DF, Dickey CC, Kasai K, Toner SK, et al. Middle and inferior temporal gyrus gray matter volume abnormalities in chronic schizophrenia: An MRI study. American Journal of Psychiatry. 2004;161(9):1603–11. WOS:000223800600013. pmid:15337650
  45. 45. Klein I, Paradis A-L, Poline J-B, Kosslyn SM, Le Bihan D. Transient Activity in the Human Calcarine Cortex During Visual-Mental Imagery: An Event-Related fMRI Study. Journal of Cognitive Neuroscience. 2000;12(Supplement 2):15–23. pmid:11506644
  46. 46. Barat E, Wirth S, Duhamel J-R. Face cells in orbitofrontal cortex represent social categories. Proceedings of the National Academy of Sciences. 2018;115(47):E11158–E67. pmid:30397122
  47. 47. Almeida I, van Asselen M, Castelo-Branco M. The role of the amygdala and the basal ganglia in visual processing of central vs. peripheral emotional content. Neuropsychologia. 2013;51(11):2120–9. https://doi.org/10.1016/j.neuropsychologia.2013.07.007.
  48. 48. Buhle JT, Kober H, Ochsner KN, Mende-Siedlecki P, Weber J, Hughes BL, et al. Common representation of pain and negative emotion in the midbrain periaqueductal gray. Social Cognitive and Affective Neuroscience. 2012;8(6):609–16. pmid:22446299
  49. 49. Sambuco N, Bradley MM, Herring DR, Lang PJ. Common circuit or paradigm shift? The functional brain in emotional scene perception and emotional imagery. Psychophysiology. 2020;57(4):e13522. pmid:32011742
  50. 50. Kober H, Barrett LF, Joseph J, Bliss-Moreau E, Lindquist K, Wager TD. Functional grouping and cortical–subcortical interactions in emotion: A meta-analysis of neuroimaging studies. NeuroImage. 2008;42(2):998–1031. pmid:18579414
  51. 51. Bzdok D, Schilbach L, Vogeley K, Schneider K, Laird AR, Langner R, et al. Parsing the neural correlates of moral cognition: ALE meta-analysis on morality, theory of mind, and empathy. Brain Structure and Function. 2012;217(4):783–96. pmid:22270812
  52. 52. Schurz M, Radua J, Aichhorn M, Richlan F, Perner J. Fractionating theory of mind: A meta-analysis of functional brain imaging studies. Neuroscience & Biobehavioral Reviews. 2014;42:9–34. pmid:24486722
  53. 53. Frith CD, Wolpert DM, Frith U, Frith CD. Development and neurophysiology of mentalizing. Philosophical Transactions of the Royal Society of London Series B: Biological Sciences. 2003;358(1431):459–73.
  54. 54. Takahashi HK, Kitada R, Sasaki AT, Kawamichi H, Okazaki S, Kochiyama T, et al. Brain networks of affective mentalizing revealed by the tear effect: The integrative role of the medial prefrontal cortex and precuneus. Neuroscience Research. 2015;101:32–43. pmid:26197267
  55. 55. Craig AD. Emotional moments across time: a possible neural basis for time perception in the anterior insula. Philosophical Transactions of the Royal Society B: Biological Sciences. 2009;364(1525):1933–42.
  56. 56. Knutson B, Greer SM. Anticipatory affect: neural correlates and consequences for choice. Philosophical Transactions of the Royal Society B: Biological Sciences. 2008;363(1511):3771–86. pmid:18829428
  57. 57. Muller-Bardorff M, Bruchmann M, Mothes-Lasch M, Zwitserlood P, Schlossmacher I, Hofmann D, et al. Early brain responses to affective faces: A simultaneous EEG-fMRI study. Neuroimage. 2018;178:660–7. WOS:000438467800055. pmid:29864521
  58. 58. Adolfi F, Couto B, Richter F, Decety J, Lopez J, Sigman M, et al. Convergence of interoception, emotion, and social cognition: A twofold fMRI meta-analysis and lesion approach. Cortex. 2017;88:124–42. Epub 20161229. pmid:28088652.
  59. 59. Kringelbach ML, Rolls ET. Neural correlates of rapid reversal learning in a simple model of human social interaction. Neuroimage. 2003;20(2):1371–83. pmid:14568506.
  60. 60. Rolls ET. The orbitofrontal cortex and emotion in health and disease, including depression. Neuropsychologia. 2019;128:14–43. Epub 20170924. pmid:28951164.
  61. 61. Ochsner KN, Bunge SA, Gross JJ, Gabrieli JD. Rethinking feelings: an FMRI study of the cognitive regulation of emotion. Journal of Cognitive Neuroscience. 2002;14(8):1215–29. pmid:12495527
  62. 62. Urry HL, van Reekum CM, Johnstone T, Davidson RJ. Individual differences in some (but not all) medial prefrontal regions reflect cognitive demand while regulating unpleasant emotion. NeuroImage. 2009;47(3):852–63. pmid:19486944
  63. 63. Picó-Pérez M, Radua J, Steward T, Menchón JM, Soriano-Mas C. Emotion regulation in mood and anxiety disorders: A meta-analysis of fMRI cognitive reappraisal studies. Progress in Neuro-Psychopharmacology and Biological Psychiatry. 2017;79:96–104. pmid:28579400
  64. 64. Davis M, Whalen PJ. The amygdala: vigilance and emotion. Molecular Psychiatry. 2001;6(1):13–34. pmid:11244481
  65. 65. Balleine BW, Killcross S. Parallel incentive processing: an integrated view of amygdala function. Trends in Neurosciences. 2006;29(5):272–9. pmid:16545468
  66. 66. Cunningham WA, Van Bavel JJ, Johnsen IR. Affective Flexibility:Evaluative Processing Goals Shape Amygdala Activity. Psychological Science. 2008;19(2):152–60. pmid:18271863.
  67. 67. Kensinger EA, Schacter DL. Processing emotional pictures and words: effects of valence and arousal. Cogn Affect Behav Neurosci. 2006;6(2):110–26. pmid:17007232.
  68. 68. Ochsner KN, Ray RD, Cooper JC, Robertson ER, Chopra S, Gabrieli JDE, et al. For better or for worse: neural systems supporting the cognitive down- and up-regulation of negative emotion. Neuroimage. 2004;23(2):483–99. WOS:000224817100005. pmid:15488398
  69. 69. Li F, Yin S, Feng P, Hu N, Ding C, Chen A. The cognitive up- and down-regulation of positive emotion: Evidence from behavior, electrophysiology, and neuroimaging. Biol Psychol. 2018;136:57–66. Epub 2018/05/23. pmid:29787789.
  70. 70. Chen Z, Whitney D. Tracking the affective state of unseen persons. PNAS Proceedings of the National Academy of Sciences of the United States of America. 2019;116(15):7559–64. pmid:30814221
  71. 71. Brosch T, Pourtois G, Sander D. The perception and categorisation of emotional stimuli: A review. Cognition and Emotion. 2010;24(3):377–400.
  72. 72. Anzellotti S, Houlihan SD, Liburd S, Saxe R. Leveraging Facial Expressions and Contextual Information to Investigate Opaque Representations of Emotions. Emotion. 2021;21(1):96–107. WOS:000614363500008. pmid:31580092