Skip to main content
Advertisement
Browse Subject Areas
?

Click through the PLOS taxonomy to find articles in your field.

For more information about PLOS Subject Areas, click here.

  • Loading metrics

Spatiotemporal dynamics of reward and punishment effects induced by associative learning

  • Huan Wang,

    Roles Conceptualization, Data curation, Formal analysis, Investigation, Methodology, Project administration, Resources, Software, Validation, Visualization, Writing – original draft, Writing – review & editing

    Affiliation Department of Psychology, Simon Fraser University, Burnaby, BC, Canada

  • Killian Kleffner,

    Roles Supervision, Visualization, Writing – review & editing

    Affiliation Department of Psychology, Simon Fraser University, Burnaby, BC, Canada

  • Patrick L. Carolan,

    Roles Software, Supervision, Writing – review & editing

    Affiliation Department of Psychology, Simon Fraser University, Burnaby, BC, Canada

  • Mario Liotti

    Roles Conceptualization, Formal analysis, Funding acquisition, Investigation, Methodology, Resources, Supervision, Writing – review & editing

    mliotti@sfu.ca

    Affiliations Department of Psychology, Simon Fraser University, Burnaby, BC, Canada, Department of Developmental and Social Psychology and Padua Neuroscience Center, University of Padua, Padua, Italy

Abstract

While reward associative learning has been studied extensively across different species, punishment avoidance learning has received far less attention. Of particular interest is how the two types of learning change perceptual processing of the learned stimuli. We designed a task that required participants to learn the association of emotionally neutral images with reward, punishment, and no incentive value outcomes through trial-and-error. During learning, participants received monetary reward, neutral outcomes or avoided punishment by correctly identifying corresponding images. Results showed an early bias in favor of learning reward associations, in the form of higher accuracy and fewer trials needed to reach learning criterion. We subsequently assessed electrophysiological learning effects with a task in which participants viewed the stimuli with no feedback or reinforcement. Critically, we found modulation of two early event-related potential components for reward images: the frontocentral P2 (170–230 ms) and the anterior N2/Early Anterior Positivity (N2/EAP; 210–310 ms). We suggest that reward associations may change stimuli detection and incentive salience as indexed by P2 and N2/EAP. We also reported, on an exploratory basis, a late negativity with frontopolar distribution enhanced by punishment images.

1. Introduction

We remember not only individuals, objects, or situations from the past, but also their associations with pleasant or unpleasant outcomes. Such associations dramatically bias our decisions towards profitable outcomes [13]. Animal research suggests that neutral cues become highly attractive when they are paired with food reward, with rats attempting to consume the inedible cue items [4, 5]. This is because pairing with a food reward confers incentive salience to the conditioned stimulus, a mechanism dependent on the dopaminergic mesolimbic system, particularly the ventral striatum [6]. Human neuroimaging studies further suggest that the orbitofrontal cortex (OFC) is also involved in reward processing [711], which has been proposed to store higher-order cognitive representations of the incentive salience [12].

In contrast, while there is evidence showing neural correlates of punishment avoidance learning in humans [1316], the mechanism through which avoiding punishment could influence the mental representation of cues and choices is not fully established. One account proposes that avoiding punishment itself can be rewarding, with the absence of an aversive outcome serving as a positive reinforcer to bias choice [14]. However, a lesion study found that damage to anterior insula and dorsal striatum selectively impaired punishment avoidance but not reward learning [15], suggesting some dissociations underlying the two types of learning. Nevertheless, at least in some cases, the OFC seems to contribute to both reward and punishment avoidance learning, which has led to the suggestion that OFC encodes an unsigned value representation for both reward and punishment cues [17, 18]. Yet, how punishment avoidance learning could change the mental representations of the predictive cues is not well understood.

Utilizing the measurement benefits of human electroencephalography (EEG) may provide another powerful way to study associative learning, as it provides a temporal characterization of exactly ‘when’ different neural processes are activated [19]. Using event-related potentials (ERPs), a human reward conditioning study found an early frontocentral component (P2, 180–280 ms) that was larger for reward than neutral cues [20]. Participants also had an enhanced P2 when anticipating monetary reward with improved task performance on reward-cued trials. These results suggest that P2 modulations may reflect initial attentional capture towards reward cues [2123]. Behavioral results have provided further support for such interpretation, showing that reward cues indeed attract attention and influence task performance, even when they are task-irrelevant [2426].

A later latency ERP component with frontal scalp distribution, the anterior N2 or N2a (180–300 ms), has been found to be modulated in its amplitude when participants anticipate monetary reward and prepare for subsequent responses [21, 23]. A number of studies have reported similar N2a modulations in response to task-irrelevant emotional stimuli, such as peripheral emotional distractors relative to neutral distractors [27,28] or emotional relative to neutral words or pictures in emotional Stroop tasks [2933]. In the latter studies, performed by our group, because emotional stimuli were consistently more positive in amplitude than neutral stimuli, we named the effect “early anterior positivity” or EAP. Given that greater EAP was found in response to drug-related images in habitual marijuana users relative to non-users [31] and in response to chocolate images in individuals who crave chocolate relative to non-cravers [32], it was proposed that enhanced EAP amplitudes indexed an early exaggerated attentional capture due to the increased motivational salience of the craved substances. Consistent with previous fMRI results (e.g., [7]), source localization of the EAP showed a main source in OFC [3132]. Altogether, these studies corroborate the claim that the EAP may constitute another candidate for an electrophysiological index of the incentive salience representation. Note that in the studies reported above the personal salience of the stimuli was acquired well before the experiments were conducted, with large inter-subject variations likely reflecting various degrees of attentional bias towards those stimuli. It is less clear how such personal salience built within a laboratory setting may be indexed by N2/EAP.

Classical conditioning studies using punishment (e.g., aversive odors or electric shock) have found mainly early visual modulations at occipital sites [34, 35]. One study found an early prefrontal modulation to aversively conditioned cues, which was proposed to be a part of the circuit that detected potential threats in the environment [36]. However, none of the studies focused on punishment avoidance learning, or compared punishment avoidance learning to reward learning. One recent ERP study has incorporated elements of both monetary gain and loss in the course of trial-and-error learning. In this task, participants learned to identify novel characters that were associated with monetary gain, loss, or neutral outcomes. Characters were presented in pairs and participants learned to pick the rewarded characters to achieve maximum payoff. In the testing phase, scalp EEG was recorded while participants made old/new judgments on learned versus novel characters. Modulations of early occipital and late parietal components were found for the rewarded characters, while no changes were found frontally or for the punishment characters [37]. Because only limited behavioral results were reported from the learning phase, it was difficult to interpret how reward learning was related to the reward ERP effects. Similarly, punishment associations were actually established through incurring monetary loss rather than avoiding such loss, which offered limited insight on punishment avoidance learning and its subsequent effects on perception.

To address the limitations of the previous studies, we designed a paradigm involving an associative learning task with both monetary reward, punishment and a neutral control condition. Critically, unlike previous studies in which the task was terminated once the learning criterion was achieved [37] or incorporated a limited number of learning trials (e.g., [38]), we set the period of learning to be sufficiently long (30 minutes) allowing participants to continue the task for the full duration even after reaching the learning criterion. This generated enough trials to allow for a thorough learning analysis. Moreover, three groups of cues were assigned with three different response buttons and strategically pressing only one button produced no long-term benefit. Following the associative learning task, all cues were presented in the context of an implicit task in which participants were instructed to attend to the physical characteristics of the cues. No feedback or monetary contingency was provided for this task. We also examined how behavioral data of learning were related to the subsequent perceptual processing of the cues.

Our first hypothesis was that an ERP modulation of the frontocentral P2 would be observed for reward cues as an index of early attentional capture. The frontal P2 is consistently reported in studies of reward anticipation [22, 23] and reward conditioning [20]. This is consistent with the notion that cues with reward incentive salience served as a “motivational magnet” attracting attention towards it [4, 6]. Secondly, we predicted that another potential ERP index of incentive salience often associated with emotional stimuli, the N2/EAP, would be enhanced for reward cues [29, 31, 39].

The last goal of the current report was to explore distinctive ERPs elicited during perceptual processing of punishment associated images following avoidance learning. No previous ERP component has been reported relating to punishment associations. Evidence from fMRI literature suggests that the OFC and anterior insula, in conjunction with the dorsal striatum, are implicated in avoidance learning and access to punishment representations [10, 14, 15]. An electrophysiological counterpart of activity generated by these brain regions is likely to yield effects over anterior frontal scalp locations, which we looked for in our data, without a clear prediction of the exact timing of such activity. We resolved to conduct an exploratory analysis on a putative punishment-related frontal component that could shed light on future research on punishment association and avoidance learning.

2. Methods

2.1. Participants

Forty participants (26 female, 32 right-handed, age mean = 20.4, range = 17–33) took part in the experiment in exchange for course credits. All provided written informed consent prior to their participation. All reported having normal or correct-to-normal vision, and none was color-blind. The Simon Fraser University Research Ethics Board approved all components of the current study.

2.2. Procedure

After providing informed consent, participants were prepared for EEG recording while they completed a demographic questionnaire. They were then transferred to a sound-attenuated room and completed the learning task, followed by the recording task. Immediately after, participants completed a subjective rating task and an affective priming task, the data of which will not be a main focus of the current report. Sessions were 2-hours long, with all tasks completed on the same day.

2.2.1. Associative learning task.

Participants viewed a serial presentation of stimuli displayed on an LCD monitor against a black background. They were informed that each stimulus belonged to one of three preassigned groups: reward, punishment, or neutral. Stimuli consisted of 18 pictures depicting mundane objects or abstract scenes (e.g., a plant, a book, a spoon, colored shapes etc.) selected from the International Affective Picture System (IAPS; [40]). The IAPS reference numbers are 5740, 7000, 7003, 7004, 7006, 7010, 7012, 7025, 7026, 7031, 7035, 7040, 7080, 7090, 7100, 7160, 7175 and 7182. Stimuli preassigned to the three groups (reward, punishment and neutral) had normative mean valence ratings of 4.85 (0.21), 5.03 (0.24), 5.12 (0.18), and arousal ratings of 2.52 (0.47), 2.74 (0.77), 2.50 (0.39). One-way analyses of variance (ANOVA) indicated that these ratings were not significantly different from one another (ps > .14). The three preassigned groups were counterbalanced across participants to minimize perceptual differences or subjective affective responses.

Stimuli were presented in pseudo-random order, with the restriction that all groups had to be sampled from once before the next iteration, and repeated for a total of 30 minutes. Participants were instructed to indicate on each trial the category they believed the stimulus belonged to by pressing the left, down, or right arrow key on a keyboard. Stimuli remained on screen until the participant responded and were followed by a feedback stimulus showing both the actual and alternative outcomes for 1000 ms. Correct responses in the reward condition earned 5 points, incorrect responses in the punishment condition lost 5 points, and no point was assigned otherwise (Fig 1A). Participants were asked to pay equal attention to all stimuli, independent of their incentive value. All points earned in the task were paid in Canadian dollars (CAD) with a conversion rate of 200:1 (M = 4 CAD). No EEG was recorded during learning due to concerns with eye blinks as well as distraction from learning.

thumbnail
Fig 1.

(A) Associative learning task: participants learned from trial-and-error which stimuli were from reward, punishment or neutral group. Each group of cues was preassigned with one button (left, down or right arrow key). Participants gained five points by correctly responding to the reward cues whereas they lost five points by incorrectly responding to the punishment cues. Zero point was given to all other cases. All points were reimbursed with Canadian dollars with a ratio of 200:1. Learning was self-paced, and lasted for 30 minutes for all participants. Feedback including the received and the alternative outcomes was shown for 1000 ms after each response. (B) Recording task: each block had 20 trials that featured the 18 previously learned stimuli from the associative learning task, two of which were repeated within each block as fillers. Participants were asked to respond to the fillers. All stimuli were shown for 500 ms, with an inter-trial interval (ITI) jittered between 1300 to 2000 ms. A fixation cross was superimposed to the center of the stimuli across the task. EEG was recorded during this task and participants were given breaks between blocks to rest and stretch. Note stimuli in the figure were for illustration only and see Methods (2.2.1.) for IAPS reference numbers.

https://doi.org/10.1371/journal.pone.0199847.g001

2.2.2. Recording task.

The 18 stimuli from the associative learning task were presented again in blocks of 20 trials. Two images were selected to be fillers and required responses from participants. Stimuli were presented for 500 ms, followed by a randomly jittered inter-trial interval of 1300 to 2000 ms. A central fixation cross was displayed throughout the task. Participants were instructed to respond to stimuli that they saw repeated a second time in that block, by pressing the key “1” on the number pad of a keyboard with their right-hand index finger (Fig 1B). Images selected as filler trials and their locations of occurrence were randomized between blocks but consistent across participants. Filler trials and response errors (i.e., false alarms to non-fillers) were dropped from ERP analyses. Participants completed 20 blocks of the task, separated by short breaks to rest.

2.3. EEG recording and processing

Continuous EEG was acquired during the recording task from 64 active Ag/AgCl electrodes based on the standard 10-10 montage (BioSemi Active Two system, Amsterdam, NL). Two channels the Common Mode Sense (CMS) and Driven Right Leg (DRL) formed a feedback loop that served as the ground. EEG was referenced to CMS during online recording without any filter. Six additional channels were applied: one at each lateral canthus (for horizontal eye movements), one below each eye (for vertical eye movements and blinks) and one over each mastoid bone. Continuous EEG was digitized at a sampling rate of 512 Hz. Semi-automatic artifact rejection was implemented using BESA with a 120 μV cut-off from 200 ms before to 800 ms after the stimulus onset. All trials were also visually inspected for additional artifacts. On average, a proportion of .34 (.19), .33 (.19) and .33 (.19) of the trials were rejected for the reward, punishment and neutral condition respectively, with no significant difference among them (p > .25). All signals were filtered offline with a high-pass filter of 0.01 Hz (zero phase, 12 dB/octave slope; BESA 5.3, MEGIS Software GmbH, Gräfelfing, Germany) and a low-pass filter of 30 Hz (zero phase, 12 dB/octave slope; FieldTrip, [41]). Bad channels were interpolated using spherical spline method implemented in FieldTrip [41]. One participant who had hairstyle not compatible with our EEG cap completed the study without EEG recorded. One participant was dropped from ERP analysis due to noisy mastoid channel data, along with two other participants for poor overall data quality producing less than 30 trials per condition. The final sample for our ERP analyses contained 36 participants. ERPs were time-locked to stimulus onset, aligned with a 200 ms prestimulus baseline, and re-referenced to an average mastoid reference. This reference system was preferred over the average reference for consistency with previous studies in our laboratory demonstrating sizeable N2/EAP effects [29, 3132].

2.4. ERP components of interest

To allow for the selection of the regions of interest (ROIs), and the latency of the effects to be independent (orthogonal) to the conditions of interest, ROIs and time windows of the ERPs were chosen on a-priori predictions based on published data and upon visual inspection of the grand-average ERPs collapsed across conditions [42]. First, the anterior P2 peaked around 200 ms on the collapsed grand-average ERPs; A ROI including frontocentral electrodes (FC1, FCz, FC2 and Cz) and a time window between 170 and 230 ms were chosen based on previous literature reporting salience-related P2 effects [20, 21, 23, 42].

Next, the anterior N2 peaked at 260 ms; a left anterior frontal ROI was selected (Fp1, AF7, AF3, F7, F5 and F3), based on previous literature highlighting the anterior N2 as modulated by biological salience [27, 33], as well as results from our lab showing a left-lateralized N2/EAP effect for personally salient appetitive stimuli [31, 32]. Time window was set between 210 and 310 ms.

Finally, we observed a later frontopolar punishment-related ERP effect evident only upon inspection of the condition-specific grand-average ERPs, in the form of a sustained larger negativity to punishment than reward and neutral stimuli, spanning from 350 ms to about 800 ms (peaking around 450 ms; named Frontopolar Negativity, or FN; Fig 2). In the absence of prior studies in the literature reporting such effect, the orthogonality requirements were not met [42]. An exploratory analysis was carried out on such punishment-related frontal effect, employing an ROI including sites Fp1, Fp2, AF3, AF4, AF7 and AF8 (the electrode locations of larger amplitude), and a time window of 350 to 650 ms. For all three ERP time windows (P2, N2/EAP, FN) mean amplitudes were extracted for statistical analyses, averaging across electrodes in the ROIs.

thumbnail
Fig 2.

(A) Grand-averaged ERP waveforms for reward, punishment and neutral stimuli (N = 36). Waveforms reflect a selected array of 18 electrodes. Positive voltages are shown below the abscissa. The epoch is shown from -200 ms to 700 ms and each tick on the abscissa represents 100 ms. (B) Grand-averaged waveforms for all external electrodes and electrodes above the eyes. IO1-2: left and right infraorbital (below the eyes), LO1-2: left and right lateral ocular (beside the eyes), M1-M2: left and right mastoid, Fp1-2: left and right frontopolar (above the eyes). Note how the peak corresponding to the FN over frontopolar sites does not invert polarity below the eyes, or between left and right lateral ocular channels.

https://doi.org/10.1371/journal.pone.0199847.g002

2.5. Analysis

2.5.1. Associative learning task.

We estimated the learning points (LP) for each condition by identifying the first trials in which participants executed at least six consecutive correct responses. We chose this criterion because there were six pictures from each condition and correctly responding to all stimuli in a particular condition was the evidence of learning. We reported the LP measure as a percentage of trials completed, i.e. the trial in which a participant reached the criterion devided by the total number of trials completed for that condition. This denominator was the same for all conditions. Reaction time (RT) and accuracy data were subgrouped into pre-learning and post-learning phases based on LP. In order to minimize the impact of outliers with our self-paced learning task, RT and accuracy data that were more than 2 SDs from the means were excluded from analyses. Cases where participants did not meet the learning criterion for a condition were dropped on an individual basis. After exclusion the final sample contained 29 subjects for RT and 27 for accuracy analyses. Mean accuracy and median RT were computed. Finally, we collected the total number of button presses for left-arrow, down-arrow and right-arrow keys across learning.

RT and accuracy data were entered into two-way ANOVA with condition and learning (pre-learning, post-learning) as within-subjects factors. Learning point and button-press data were entered into one-way ANOVA with condition as within-subjects factor.

2.5.2. Recording task.

For each participant and condition mean accuracy was computed separately for hits as well as false alarms. Sensitivity index d-Prime was calculated by subtracting z-transformed false alarm score from hits z-score. For each participant, ERP mean amplitudes (P2, N2/EAP, FN windows) were entered into one-way ANOVAs with condition as within-subject factor. For all ANOVAs (behavioral and ERP data), post hoc t-tests used the Bonferroni family-wise error correction to assess significant main effects and interactions. Alpha level was set to .05 and the Greenhouse-Geisser correction was applied when the assumption of sphericity was violated.

3. Results

3.1. Behavioral effects

3.1.1. Associative learning task.

The assumption of sphericity was violated (Mauchly’s test p = .003) and we reported adjusted degree of freedom for this ANOVA. The ANOVA of the learning point estimation returned a main effect of condition (F (1.6, 56.31) = 10.84, p < .001, ƞ2p = .23), with reward cues being learned faster than punishment and neutral cues (t (36) = -2.59, p = .046; t (36) = -4.42, p < .001, respectively; see Table 1). There was also a trend for faster learning of punishment over neutral associations (t (36) = -2.12, p = .12). The median number of trials participants saw per condition was 245, with a range of 142 to 336 trials.

Button presses also showed a main effect of condition (F (2, 78) = 10.61, p < .001, ƞ2p = .21), with reward and punishment buttons being pressed more than neutral buttons (t (39) = 4.25, p < .001; t (39) = 2.67, p = .033, respectively; Table 1).

Accuracy analysis showed a significant condition x learning two-way interaction (F (2, 52) = 7.00, p = .002, ƞ2p = .21). It was explained by a main effect of condition before learning (F (2, 52) = 11.31, p < .001, ƞ2p = .30; reward > punishment, t (26) = 4.09, p = .001, reward > neutral, t (35) = 3.47, p = .006). In contrast, only trend level accurary difference was detected after learning (F (2, 52) = 3.29, p = .045, ƞ2p = .11; reward > punishment, t (26) = 2.47, p = .061, reward > neutral, t (35) = 1.94, p = .19). These results suggested that participants showed an early learning bias towards reward cues, making more attempts towards receiving rewards.

The two-way ANOVA on split RT only returned a main effect of learning (F (1, 28) = 71.4, p < .001, ƞ2p = .72), with faster RT after (M = 769.7 ms, SD = 23.3) than before learning (M = 1764.2 ms, SD = 119.2). No interaction or main effect of condition was observed (ps > .22).

3.1.2. Recording task.

No main effects were found for hit or false alarm rates (p > .26). Similarly, no d-prime difference was detected (p = .37; reward M = 1.49, SD = .67, punishment M = 1.31 SD = 1.05, neutral M = 1.42, SD = .78). Collapsing across conditions, the data showed that participants followed the instructions well (false alarm M = .032, SD = .037) and remained engaged in the task (hit M = .54, SD = .19).

3.2. ERP effects

3.2.1. P2 mean amplitude (170–230 ms).

P2 amplitude showed a condition dependent difference (F (2, 70) = 3.23, p = .046, ƞ2p = .084), with reward trials (M = -2.66 μV, SD = 4.11) having greater amplitude than punishment trials (M = -3.44 μV, SD = 4.45; t (35) = 2.70, p = .032, d = 0.18). The other two comparisons did not reach significance (reward vs. neutral, t (35) = 1.91, p = .19; punishment vs. neutral, t (35) = -.56, p = 1.0).

3.2.2. N2/EAP mean amplitude (210–310 ms).

Similarly, for the N2/EAP amplitude there was a main effect of condition (F (2, 70) = 3.63, p = .032, ƞ2p = .094), due to a significant difference between the reward (M = -4.39 μV, SD = 3.72) and punishment conditions (M = -5.35 μV, SD = 3.56; t (35) = 2.88, p = .020, d = 0.27). The other two comparisons did not reach significance (reward vs. neutral, t (35) = 1.71, p = .29; punishment vs. neutral, t (35) = -.88, p = 1.0).

3.2.3. FN mean amplitude (350–650 ms).

Finally, we found a main effect of condition for FN (F (2, 70) = 4.58, p = .014, ƞ2p = .12), as a result of difference between the punishment (M = -5.12 μV, SD = 4.83) and reward conditions (M = -3.78 μV, SD = 4.50; t (35) = -2.65, p = .036, d = 0.29), along with a trend level difference between punishment and neutral condition (M = -3.94 μV, SD = 3.93; t (35) = -2.26, p = .091, d = 0.27). Bear in mind, however, that the FN analysis was exploratory in nature and should be interpreted with caution.

3.3. Correlations to behavioral effects

Correlational analyses were conducted using Spearman’s rank-order test to account for significant outliers in our data. In light of our accuracy results, we looked at how pre-learning accuracy difference was related to P2 and N2/EAP difference waves of the significant contrasts (reward minus punishment). No difference was found (ps > .67).

We further explored more informative measures looking at how specific incorrect choices for each condition may impact learning. For example, when participants made errors for reward cues, they would make either punishment or neutral judgments. We speculated that pressing punishment button to reward cues might create a larger impact on learning and representations of the reward cues than making a neutral judgment. The same logic would apply to reward judgments towards punishment cues. Therefore, we calculated the difference score between the ratios of punishment errors on reward trials (i.e., updating reward associations) and reward errors on punishment trials (reinforcement error difference). A higher value may suggest a stronger bias towards reward over punishment learning. Furthermore, with the inclusion of neutral cues, we calculated the difference between ratios of reward and punishment errors for neutral cues. This measure may suggest individual differences in pursuing reward versus avoiding punishment under uncertainty independent of reward or punishment learning, given that errors on neutral trials only informed learning of neutral associations and incurred no monetary outcome. We further controlled the general tendency towards choosing one outcome by correcting for the number of times each button was pressed. Hence, reward and punishment errors on neutral trials were divided by the total number of times each button was pressed before the difference was calculated (i.e., reward errors / reward BP—punishment errors / punishment BP). A larger value for this measure (neutral error difference) may suggest higher reward over punishment sensitivity. We correlated these two behavioral measures with P2 and N2/EAP which showed significant contrasts between reward and punishment conditions.

Furthermore, we correlated these two measures with difference scores (reward minus punishment) of pre and post-learning RT and ACC as control analyses: if they indeed tap into learning, reward and punishment sensitivity, we would expect them to be related to RT or ACC during learning. We found that reinforcement error difference (M = -.045, SD = .260) had a trend positive relationship with pre-learning ACC difference (rs = .31, p = .059). In contrast, neutral error difference (M = -.00064, SD = .0011) was related to post-learning ACC difference although in the negative direction (rs = -.59, p < .001).

Interestingly, larger reinforcement error difference was correlated with higher P2 difference (rs = .39, p = .018; see Fig 3), but not N2/EAP (rs = .18, p = .29). In contrast, larger neutral error difference was correlated with both higher N2/EAP effect (rs = .38, p = .023), and P2 effect (rs = .38, p = .026). Nevertheless, we note that these correlation analyses were reported as exploratory and many will not survive correction for multiple comparisons.

thumbnail
Fig 3.

Topographical maps for the three components observed for the current study. All topographical maps are based on difference waves: P2, EAP (reward—punishment), FN (punishment—reward). Note how the N2/EAP and FN do not invert polarity below or between left and right ocular channels, suggesting the source of these activities is neural rather than ocular artifacts. Scatter plots show correlations between P2 contrast and reinforcement error difference (punishment errors on reward trials minus reward errors on punishment trials), and between N2/EAP contrast and neutral error difference (reward minus punishment errors at neutral trials).

https://doi.org/10.1371/journal.pone.0199847.g003

4. Discussion

In the current study, participants learned to associate uninteresting images with reward, punishment or neutral monetary outcomes. We observed an early bias towards learning reward associations, although all associations were eventually learned. Brain electrophysiological responses to visual presentations of the images during a subsequent implicit task showed differential spatiotemporal effects. Specifically, when images were associated with monetary reward, earlier stages of stimulus detection and evaluation were affected (P2 and N2/EAP; 170–310 ms). In contrast, we observed a later post-evaluative effect for images associated with monetary punishment (FN; 350–650 ms).

4.1. Behavioral effects of associative learning

Even though participants made more attempts towards receiving rewards and avoiding punishments, as indicated by the larger number of reward and punishment button-press than neutral, accuracy and learning trials data suggested that reward associations, but not punishment associations, were preferably learned. One explanation for the different learning profiles is that reward and punishment may act in opposite manner when interacting with behavioral control systems, with reward encourages active approach and punishment facilitates behavioral inhibition [43, 44]. For example, when reward and punishment were orthogonally manipulated with action (button-press versus button-release) in a go/no-go task, Guitart-Masip and colleagues [45] found that reward facilitated learning go responses whereas punishment motivated more no-go responses. Given that our learning task is a go task, our results were consistent with this framework of valence and action interaction. In contrast, when comparing learning and response biases for punishment and neutral conditions, we observed a different profile: punishment cues were learned faster than neutral cues (p = .039, uncorrected) and punishment buttons received more preference over neutral buttons. Therefore, we further suggest that the interaction between valence and action seems to influence specific strategies during learning which prompts biases towards reward learning. However, valence (and salience) by itself, positive or negative, also seems to promote learning.

Our accuracy results are in line with previous studies using reward cues to motivate performance, showing that participants are more accurate for high reward trials than low reward or neutral trials [4648]. We demonstrate here that such reward motivation can be present even when participants simultaneously learned to avoid monetary punishment. However, we did not observe any RT difference reported previously. This could be a result of task instruction: our learning task was self-paced rather than timed, which emphasized accuracy over speed. Therefore, participants may have been motivated to maintain high accuracy rather than faster RT.

4.2. ERP components related to associative learning

4.2.1. P2 wave.

Our first hypothesis, that the frontocentral P2 elicited by rewarded stimuli would be enhanced was supported. This result is consistent with paradigms using monetary reward to motivate task performance. In such tasks, enhanced P2 to reward cues was interpreted as an indication of motivated stimulus processing to obtain the impending reward [22]. The current result is also consistent with the interpretation that the P2 effect is larger for passively viewed reward-predictive cues due to their increased salience [20, 26], as well as P2 as an index of automatic attentional capture [28]. However, it has been difficult to discern whether P2 was produced by active reward anticipation, learned incentive salience, or attentional capture when EEG recording takes place during reward learning. In the present study, P2 was elicited after reward learning during an implicit task in which no form of reinforcement was provided. Therefore, the current study may suggest the possibility that P2 modulations could take place implicitly. Interestingly, making more punishment errors during reward trials cues (i.e., updating reward associations) and less reward errors for punishment cues were correlated with larger P2 effects. Presumably, such an asymmetry further shapes learning and may contribute to later observed P2 difference. Future research utilizing trial to trial analyses could test this possibility.

Other ERP components share a similar frontocentral scalp distribution to the P2, albeit later in time (200–300 ms). These include the feedback-related negativity (FRN; [49, 50]) and the reward positivity (RP; [51]). Both components are elicited by feedback and feedback-predicting cues, with FRN greater for monetary loss and RP greater for monetary gain [52, 53]. It is an ongoing debate whether FRN/RP modulations during this time interval are driven by negative deflections to punishment or positive changes to reward cues [51, 54, 55]. We believe that the P2 effect found in the current study differs from FRN/RP effects reported in previous studies. First, our P2 emerged and peaked earlier (170–230 ms) than the FRN/RP, indicating a temporal dissociation. Additional support for our claim comes from previous functional interpretations of the P2 as stimuli-based attentional modulation whereas FRN was linked to action monitoring or controlled attention [26, 28]. In this regard, the FRN/RP are better associated with the frontocentral N2. This component is thought to play a role in conflict monitoring [51, 56], and produces the baseline for FRN/RP modulations [55]. The frontocentral N2 also is distinguished from the P2 in more traditional visual and auditory attention tasks [57, 58].

We agree with the functional interpretation of P2 and FRN/RP. Importantly, in our recording task participants simply viewed the previously learned cues without any current reinforcement feedback or overt response. It is not surprising that without error feedback or conflict, little activity from the FRN/RP latency was observed in frontocentral sites [59]. The P2, however, might reflect stimulus-specific detection and could remain as long as participants see stimuli previously associated with reward, punishment or neutral outcomes. Interestingly, we did observe activity from the N2 time window, which is discussed in the following section.

4.2.2. N2/EAP.

Our second hypothesis that reward cues would show enhanced N2/EAP modulation (210–310 ms) was also supported. The present finding is consistent with previous studies in our laboratory employing appetitive stimuli with heightened incentive salience among certain groups of individuals—such as chocolate and marijuana cues viewed by chocolate cravers and individuals with marijuana dependence [31, 32]. Because such individuals have extensive experience with their craved substance, associated cues are suggested to be attributed with incentive salience as a result of dopaminergic activity [4, 6, 60]. It is also consistent with past studies employing aversive but salient distractors (e.g., spiders, threat words) to capture attention and influence performance [29, 33]. The N2/EAP effect in previous studies has been discussed as an index of automatic attentional capture based on fast visual sampling of significant features [31, 33, 39]. However, salience of the stimuli employed in previous research was likely developed outside of laboratory, either through real-life (marijuana) or evolution (spiders). In the present experiment, we have shown that elevated salience, indexed by the N2/EAP, can be induced with a learning session as short as thirty minutes. Moreover, the N2/EAP (and P2) effect seems to be related to individual differences in reward sensitivity. Individuals who are more reward seeking (making more reward than punishment errors) during neutral trials tend to show larger N2/EAP modulation. Note that reward or punishment errors in this case only inform learning of neutral cues, which did not factor in the observed N2/EAP difference. Given this observation, future research should test individual difference in reward-seeking and punishment avoidance traits and examine how these traits could modulate feedback-related activity during learning and the P2 and N2/EAP after learning.

Unlike the previous study [37], we did not observe any early difference in the occipital site (called very early effects of emotion; VEEE, see Fig 2). There are two plausible explanations for such discrepancy. First, ERPs in the present study were referenced to the averaged mastoids, whereas VEEE was reported with the average reference, making direct comparison difficult. A second interpretation for the present result stems from the established anatomy of the ascending dopaminergic projections of the reward system, which target frontal cortex locations far more extensively than occipitoparietal ones [61]. This should logically result in more anterior than posterior scalp distribution of reward-dependent electrophysiological activity. Yet, associating reward with highly visually challenging stimuli (i.e., Chinese characters) may selectively enhance early visual processing, establishing a more posterior effect (VEEE; [37]).

4.2.3. FN.

We conducted exploratory analyses on an observed frontal ERP effect. A noticeable negative peak with frontopolar distribution (FN) was evident particularly for the punishment condition. A similar FN effect was previously reported in response to failed versus successful inhibitions in a stop-signal task, and was interpreted as reflecting a post-evaluative error correction mechanism [62]. If such interpretation were to apply to our study, a post-evaluative error mechanism may be implicitly activated possibly in the form of punishment associations for trials not requiring an overt categorization or selection and execution of a motor response. Another study reported a similar, albeit later and more sustained frontal negativity (named frontal slow wave; 500–1500 ms) during a blackjack task in response to trials associated to losses compared to wins and ties [63]. A frontal slow wave was also reported during a counting Stroop task. Interestingly, such effects were more pronounced among individuals who experienced the greatest reduction in positive affect ratings [64], perhaps suggesting a similar sensitivity of the frontal negativity to the impact of negative affect on cognitive control operations.

Finally, another frontal component argued to be the index of familiarity-driven recognition (FN400; e.g., [65]) showed somewhat similar temporal and spatial distributions as the current FN. It could be that in our study the FN modulation reflected implicit recognition of the learned images. However, in the ERP memory literature FN400 is more positive for remembered (old) compared to new stimuli, and this effect is enhanced for images with negative emotional valence [66]. This is at odds with our findings for the FN in the present study, which was more negative in amplitude for punishment relative to reward images. Nevertheless, future research could draw from this exploratory finding, along with previous results, further demonstrating the role of this frontal negativity in the interplay between affect, error correction and cognitive control during punishment avoidance learning.

4.3. Conclusions and limitations

A number of limitations can be identified in the present study. First, our learning point estimation did not take into account potential response biases during interleaved reward and punishment avoidance learning. For example, according to our estimation procedure, a person pressing reward button across all trials would “learn” reward associations at trial one and never learn others. This would not be an accurate estimation of individual’s learning. Future research employing statistical modeling to analyze learning data is needed to address any response bias during interleaved learning (e.g., [67, 68]). Second, we did not collect EEG data during the associative learning task, which prevented us from further analyzing how learning contributes to later differences in perceptual processing. In a similar vein, we only found significant differences of reward and punishment conditions for our ERP results, but not neutral condition. One future direction is to include EEG data during or even before learning, from which one might see significant contrast against neutral condition by computing the pre and post-learning ERP difference. Another future extension of the present study could assess the longevity of reward or punishment associations as indexed by the P2, N2/EAP and FN effects reported here. Additionally, although we interpreted our results in relation to associations of reward, punishment and neutral outcomes, we could not rule out effects of perceptual or semantic priming. Indeed, participants may be primed to show differential electrophysiological responses towards our stimuli. Future studies could test this possibility by using the same pictures during learning, and record electrophysiological responses to semantically related words.

In conclusion, using an associative learning task, we found distinct trajectories of learning marked by early preference towards learning reward associations, when there was a need for competition of attentional resources. Despite an absence of behavioral differences among cues at the end of the learning period, we found larger early ERP components (P2 & N2/EAP) for reward over punishment cues. We interpreted them as indices of heightened incentive salience. Additionally, we reported an exploratory finding of a frontal negativity (FN) that appeared to be more prominent for punishment over reward cues. Future research could test the possibility of this frontal negativity involved in mediating how an aversive state can motivate additional needs for control and learning.

Acknowledgments

We thank Megan Liau and research assistants in our lab for the help with data collection. We are particularly grateful to John McDonald and Anne Smith for valuable advice on the project. This research was supported by grants from Natural Sciences and Engineering Research Council of Canada (NSERC) and Canada Foundation for Innovation (CFI) to ML.

References

  1. 1. Bechara A, Damasio H, Damasio AR. Emotion, decision making and the orbitofrontal cortex. Cereb Cortex. 2000;10(3):295–307. pmid:10731224
  2. 2. Bechara A, Dolan S, Hindes A. Decision-making and addiction (part II): myopia for the future or hypersensitivity to reward? Neuropsychologia. 2002;40(10):1690–705. pmid:11992657
  3. 3. Damasio AR. The somatic marker hypothesis and the possible functions of the prefrontal cortex. Philos Trans R Soc Lond B Biol Sci. 1996;351(1346):1413–20. pmid:8941953
  4. 4. Berridge KC. From prediction error to incentive salience: mesolimbic computation of reward motivation. Eur J Neurosci. 2012;35(7):1124–43. pmid:22487042
  5. 5. Saunders BT, Robinson TE. A cocaine cue acts as an incentive stimulus in some but not others: implications for addiction. Biological psychiatry. 2010;67(8):730–6. pmid:20045508
  6. 6. Robinson TE, Berridge KC. The neural basis of drug craving: an incentive-sensitization theory of addiction. Brain Res Brain Res Rev. 1993;18(3):247–91. pmid:8401595
  7. 7. Kahnt T, Heinzle J, Park SQ, Haynes JD. The neural code of reward anticipation in human orbitofrontal cortex. Proc Natl Acad Sci U S A. 2010;107(13):6010–5. pmid:20231475
  8. 8. Knutson B, Fong GW, Adams CM, Varner JL, Hommer D. Dissociation of reward anticipation and outcome with event-related fMRI. Neuroreport. 2001;12(17):3683–7. pmid:11726774
  9. 9. Knutson B, Fong GW, Bennett SM, Adams CM, Hommer D. A region of mesial prefrontal cortex tracks monetarily rewarding outcomes: characterization with rapid event-related fMRI. Neuroimage. 2003;18(2):263–72. pmid:12595181
  10. 10. O’Doherty J, Kringelbach ML, Rolls ET, Hornak J, Andrews C. Abstract reward and punishment representations in the human orbitofrontal cortex. Nat Neurosci. 2001;4(1):95–102. pmid:11135651
  11. 11. Ursu S, Carter CS. Outcome representations, counterfactual comparisons and the human orbitofrontal cortex: implications for neuroimaging studies of decision-making. Brain Res Cogn Brain Res. 2005;23(1):51–60. pmid:15795133
  12. 12. Berridge KC, Robinson TE. Parsing reward. Trends Neurosci. 2003;26(9):507–13. pmid:12948663
  13. 13. Delgado MR, Jou RL, Ledoux JE, Phelps EA. Avoiding negative outcomes: tracking the mechanisms of avoidance learning in humans during fear conditioning. Front Behav Neurosci. 2009;3:33. pmid:19847311
  14. 14. Kim H, Shimojo S, O’Doherty JP. Is avoiding an aversive outcome rewarding? Neural substrates of avoidance learning in the human brain. PLoS biology. 2006;4(8):e233. pmid:16802856
  15. 15. Palminteri S, Justo D, Jauffret C, Pavlicek B, Dauta A, Delmaire C, et al. Critical roles for anterior insula and dorsal striatum in punishment-based avoidance learning. Neuron. 2012;76(5):998–1009. pmid:23217747
  16. 16. Samanez-Larkin GR, Hollon NG, Carstensen LL, Knutson B. Individual differences in insular sensitivity during loss anticipation predict avoidance learning. Psychol Sci. 2008;19(4):320–3. pmid:18399882
  17. 17. Kahnt T, Park SQ, Haynes JD, Tobler PN. Disentangling neural representations of value and salience in the human brain. Proc Natl Acad Sci U S A. 2014;111(13):5000–5. pmid:24639493
  18. 18. Metereau E, Dreher J-C. The medial orbitofrontal cortex encodes a general unsigned value signal during anticipation of both appetitive and aversive events. Cortex. 2015;63:42–54. pmid:25243988
  19. 19. Tucker DM, Liotti M, Potts GF, Russell GS, Posner MI. Spatiotemporal analysis of brain electrical fields. Human Brain Mapping. 1994;1(2):134–52.
  20. 20. Franken IH, Huijding J, Nijs IM, van Strien JW. Electrophysiology of appetitive taste and appetitive taste conditioning in humans. Biol Psychol. 2011;86(3):273–8. pmid:21187120
  21. 21. Krebs RM, Boehler CN, Appelbaum LG, Woldorff MG. Reward associations reduce behavioral interference by changing the temporal dynamics of conflict processing. PLoS One. 2013;8(1):e53894. pmid:23326530
  22. 22. Schevernels H, Krebs RM, Santens P, Woldorff MG, Boehler CN. Task preparation processes related to reward prediction precede those related to task-difficulty expectation. NeuroImage. 2014;84:639–47. pmid:24064071
  23. 23. van den Berg B, Krebs RM, Lorist MM, Woldorff MG. Utilization of reward-prospect enhances preparatory attention and reduces stimulus conflict. Cogn Affect Behav Neurosci. 2014;14(2):561–77. pmid:24820263
  24. 24. Anderson BA, Laurent PA, Yantis S. Value-driven attentional capture. Proc Natl Acad Sci U S A. 2011;108(25):10367–71. pmid:21646524
  25. 25. Hickey C, Chelazzi L, Theeuwes J. Reward changes salience in human vision via the anterior cingulate. J Neurosci. 2010;30(33):11096–103. pmid:20720117
  26. 26. Potts GF, Martin LE, Burton P, Montague PR. When things are better or worse than expected: the medial frontal cortex and the allocation of processing resources. Journal of cognitive neuroscience. 2006;18(7):1112–9. pmid:16839285
  27. 27. Carretie L. Exogenous (automatic) attention to emotional stimuli: a review. Cogn Affect Behav Neurosci. 2014;14(4):1228–58. pmid:24683062
  28. 28. Carretie L, Albert J, Lopez-Martin S, Hoyos S, Kessel D, Tapia M, et al. Differential neural mechanisms underlying exogenous attention to peripheral and central distracters. Neuropsychologia. 2013;51(10):1838–47. pmid:23816957
  29. 29. Carretie L, Hinojosa JA, Martin-Loeches M, Mercado F, Tapia M. Automatic attention to emotional stimuli: neural correlates. Hum Brain Mapp. 2004;22(4):290–9. pmid:15202107
  30. 30. Taake I, Jaspers-Fayer F, Liotti M. Early frontal responses elicited by physical threat words in an emotional Stroop task: Modulation by anxiety sensitivity. Biol Psychol. 2009;81(1):48–57. pmid:19428968
  31. 31. Williams LM, Palmer D, Liddell BJ, Song L, Gordon E. The ‘when’and ‘where’of perceiving signals of threat versus non-threat. Neuroimage. 2006;31(1):458–67. pmid:16460966
  32. 32. Asmaro D, Carolan PL, Liotti M. Electrophysiological evidence of early attentional bias to drug-related pictures in chronic cannabis users. Addict Behav. 2014;39(1):114–21. pmid:24126204
  33. 33. Asmaro D, Jaspers-Fayer F, Sramko V, Taake I, Carolan P, Liotti M. Spatiotemporal dynamics of the hedonic processing of chocolate images in individuals with and without trait chocolate craving. Appetite. 2012;58(3):790–9. pmid:22326680
  34. 34. Morel S, Beaucousin V, Perrin M, George N. Very early modulation of brain responses to neutral faces by a single prior association with an emotional context: evidence from MEG. Neuroimage. 2012;61(4):1461–70. pmid:22525875
  35. 35. Steinberg C, Brockelmann AK, Rehbein M, Dobel C, Junghofer M. Rapid and highly resolving associative affective learning: convergent electro- and magnetoencephalographic evidence from vision and audition. Biol Psychol. 2013;92(3):526–40. pmid:23481617
  36. 36. Hintze P, Junghofer M, Bruchmann M. Evidence for rapid prefrontal emotional evaluation from visual evoked responses to conditioned gratings. Biol Psychol. 2014;99:125–36. pmid:24699136
  37. 37. Schacht A, Adler N, Chen P, Guo T, Sommer W. Association with positive outcome induces early effects in event-related brain potentials. Biol Psychol. 2012;89(1):130–6. pmid:22027086
  38. 38. Kahnt T, Heinzle J, Park SQ, Haynes JD. Decoding the formation of reward predictions across learning. J Neurosci. 2011;31(41):14624–30. pmid:21994378
  39. 39. Carolan PL, Jaspers-Fayer F, Asmaro DT, Douglas KS, Liotti M. Electrophysiology of blunted emotional bias in psychopathic personality. Psychophysiology. 2014;51(1):36–41. pmid:24016313
  40. 40. Lang PJ, Bradley MM, Cuthbert BN. International affective picture system (IAPS): Technical manual and affective ratings. NIMH Center for the Study of Emotion and Attention. 1997:39–58.
  41. 41. Oostenveld R, Fries P, Maris E, Schoffelen JM. FieldTrip: Open source software for advanced analysis of MEG, EEG, and invasive electrophysiological data. Comput Intell Neurosci. 2011;2011:156869. pmid:21253357
  42. 42. Luck SJ, Gaspelin N. How to get statistically significant effects in any ERP experiment (and why you shouldn’t). Psychophysiology. 2017;54(1):146–57. pmid:28000253
  43. 43. Frank MJ, Seeberger LC, O’Reilly RC. By carrot or by stick: cognitive reinforcement learning in parkinsonism. Science. 2004;306(5703):1940–3. pmid:15528409
  44. 44. Boureau Y-L, Dayan P. Opponency revisited: competition and cooperation between dopamine and serotonin. Neuropsychopharmacology. 2011;36(1):74. pmid:20881948
  45. 45. Guitart-Masip M, Huys QJ, Fuentemilla L, Dayan P, Duzel E, Dolan RJ. Go and no-go learning in reward and punishment: interactions between affect and effect. Neuroimage. 2012;62(1):154–66. pmid:22548809
  46. 46. Broyd SJ, Richards HJ, Helps SK, Chronaki G, Bamford S, Sonuga-Barke EJ. An electrophysiological monetary incentive delay (e-MID) task: a way to decompose the different components of neural response to positive and negative monetary reinforcement. J Neurosci Methods. 2012;209(1):40–9. pmid:22659003
  47. 47. Santesso DL, Bogdan R, Birk JL, Goetz EL, Holmes AJ, Pizzagalli DA. Neural responses to negative feedback are related to negative emotionality in healthy adults. Soc Cogn Affect Neurosci. 2012;7(7):794–803. pmid:21917847
  48. 48. Pfabigan DM, Seidel E-M, Sladky R, Hahn A, Paul K, Grahl A, et al. P300 amplitude variation is related to ventral striatum BOLD response during gain and loss anticipation: An EEG and fMRI experiment. NeuroImage. 2014;96:12–21. pmid:24718288
  49. 49. Gehring WJ, Willoughby AR. The medial frontal cortex and the rapid processing of monetary gains and losses. Science. 2002;295(5563):2279–82. pmid:11910116
  50. 50. Oliveira FT, McDonald JJ, Goodman D. Performance monitoring in the anterior cingulate is not all error related: expectancy deviation and the representation of action-outcome associations. Journal of cognitive neuroscience. 2007;19(12):1994–2004. pmid:17892382
  51. 51. Baker TE, Holroyd CB. Dissociated roles of the anterior cingulate cortex in reward and conflict processing as revealed by the feedback error-related negativity and N200. Biological psychology. 2011;87(1):25–34. pmid:21295109
  52. 52. Yeung N, Holroyd CB, Cohen JD. ERP correlates of feedback and reward processing in the presence and absence of response choice. Cereb Cortex. 2005;15(5):535–44. pmid:15319308
  53. 53. Holroyd CB, Krigolson OE, Lee S. Reward positivity elicited by predictive cues. Neuroreport. 2011;22(5):249–52. pmid:21386699
  54. 54. Hauser TU, Iannaccone R, Stämpfli P, Drechsler R, Brandeis D, Walitza S, et al. The feedback-related negativity (FRN) revisited: new insights into the localization, meaning and network organization. Neuroimage. 2014;84:159–68. pmid:23973408
  55. 55. Proudfit GH. The reward positivity: from basic research on reward to a biomarker for depression. Psychophysiology. 2015;52(4):449–59. pmid:25327938
  56. 56. Van Veen V, Carter CS. The timing of action-monitoring processes in the anterior cingulate cortex. J Cogn Neurosci. 2002;14(4):593–602. pmid:12126500
  57. 57. Luck SJ, Hillyard SA. Electrophysiological correlates of feature analysis during visual search. Psychophysiology. 1994;31(3):291–308. pmid:8008793
  58. 58. Potts GF, Liotti M, Tucker DM, Posner MI. Frontal and inferior temporal cortical activity in visual target detection: Evidence from high spatially sampled event-related potentials. Brain Topography. 1996;9(1):3–14.
  59. 59. Holroyd CB, Coles MGH. The neural basis of human error processing: reinforcement learning, dopamine, and the error-related negativity. Psychol Rev. 2002;109(4):679–709. pmid:12374324
  60. 60. Schultz W. Updating dopamine reward signals. Curr Opin Neurobiol. 2013;23(2):229–38. pmid:23267662
  61. 61. Delgado MR. Reward-related responses in the human striatum. Ann N Y Acad Sci. 2007;1104:70–88. pmid:17344522
  62. 62. Schmajuk M, Liotti M, Busse L, Woldorff MG. Electrophysiological activity underlying inhibitory control processes in normal adults. Neuropsychologia. 2006;44(3):384–95. pmid:16095637
  63. 63. West R, Bailey K, Tiernan BN, Boonsuk W, Gilbert S. The temporal dynamics of medial and lateral frontal neural activity related to proactive cognitive control. Neuropsychologia. 2012;50(14):3450–60. pmid:23085124
  64. 64. West R, Travers S. Tracking the temporal dynamics of updating cognitive control: an examination of error processing. Cereb Cortex. 2008;18(5):1112–24. pmid:17716989
  65. 65. Bergstrom ZM, Williams DG, Bhula M, Sharma D. Unintentional and Intentional Recognition Rely on Dissociable Neurocognitive Mechanisms. J Cogn Neurosci. 2016;28(11):1838–48. pmid:27417203
  66. 66. Schaefer A, Pottage CL, Rickart AJ. Electrophysiological correlates of remembering emotional pictures. Neuroimage. 2011;54(1):714–24. pmid:20650320
  67. 67. Smith AC, Wirth S, Suzuki WA, Brown EN. Bayesian analysis of interleaved learning and response bias in behavioral experiments. J Neurophysiol. 2007;97(3):2516–24. pmid:17182907
  68. 68. Solomon M, Smith AC, Frank MJ, Ly S, Carter CS. Probabilistic reinforcement learning in adults with autism spectrum disorders. Autism Res. 2011;4(2):109–20. pmid:21425243