Implication of the locus coeruleus in effort processing
-
Upload
khangminh22 -
Category
Documents
-
view
0 -
download
0
Transcript of Implication of the locus coeruleus in effort processing
HAL Id: tel-03484040https://tel.archives-ouvertes.fr/tel-03484040
Submitted on 16 Dec 2021
HAL is a multi-disciplinary open accessarchive for the deposit and dissemination of sci-entific research documents, whether they are pub-lished or not. The documents may come fromteaching and research institutions in France orabroad, or from public or private research centers.
L’archive ouverte pluridisciplinaire HAL, estdestinée au dépôt et à la diffusion de documentsscientifiques de niveau recherche, publiés ou non,émanant des établissements d’enseignement et derecherche français ou étrangers, des laboratoirespublics ou privés.
Implication of the locus coeruleus in effort processing :electrophysiological studies
Pauline Bornert
To cite this version:Pauline Bornert. Implication of the locus coeruleus in effort processing : electrophysiological studies.Neurons and Cognition [q-bio.NC]. Sorbonne Université, 2021. English. �NNT : 2021SORUS163�.�tel-03484040�
Sorbonne Université
École doctorale Cerveau Cognition Comportement
Institut du Cerveau et de la Moelle épinière / Équipe Motivation Cerveau et Comportement
Implication of the locus coeruleus in effort processing:
electrophysiological studies
Par Pauline Bornert
Thèse de doctorat de Neurosciences
Dirigée par Sébastien Bouret
Présentée et soutenue publiquement le 14 septembre 2021
Devant un jury composé de: Sébastien Bouret, directeur de thèse
Mark Walton, rapporteur
Etienne Coutureau, raporteur
Philippe Faure, examinateur
Fadila Hadj-Bouziane, examinatrice
Sabrina Ravel, examinatrice
1
Remerciements Tout d’abord, je souhaiterais remercier tous les membres de mon jury de thèse : Mark
Walton, Etinenne Coutureau, Philippe Faure, Fadila Hadj-Bouziane et Sabrina Ravel. Un grand merci à Philippe Faure et Fadila Hadj-Bouziane qui ont suivi mon travail au cours de mes deux CST, et m’ont prodigué des conseils précieux. Un grand merci aussi à Mark Walton et Etienne Coutureau, qui ont accepté le rôle laborieux de raporteurs.
Je souhaite aussi adresser mes remerciements à l’école doctorale Cerveau, Cognition et Comportement qui a accepté de financer mon travail.
Un grand merci à Sébastien Bouret, mon directeur de thèse. Merci de m’avoir accordé ta confiance pour mon stage de Master 2, et merci de ton soutien pendant cette année compliquée au cours de laquelle je ne pouvais pratiquement pas marcher. Merci de m’avoir renouvelé ta confiance pour mon doctorat. Merci pour ton accompagnement et ton encadrement tout au long de cette thèse, au cours de laquelle tu m’as toujours poussée à me remettre en question et à me dépasser, et pour ces quatre années de discussions scientifiques enrichissantes.
Je souhaiterais aussi remercier tous les membres de notre équipe Motivation, Brain and Behavior, présents et passés, que j’ai pu cotoyer, pour leur sympathie et la qualité des échanges scientifiques que nous avons pu avoir au cours des réunions de labo et dans l’open space. Merci en particulier à Jean et Mathias pour leurs critiques constructives lors de mes présentations.
Un grand merci à tout le personnel de la plateforme primates de l’ICM, sans qui les études présentées dans ce manuscrit n’auraient pas pu avoir lieu. Je tiens en particulier à remercier Estelle, pour sa bonne humeur et son efficacité redoutable. Un grand merci pour ton aide précieuse avec les singes, et pour les rigolades dans les vestiaires ou les couloirs. Un grand merci aussi à Sophie, Audrey, Marie et Coraline pour tout ce qu’elles ont fait ou font pour nos macaques au jour le jour.
Sur le plan personnel, je souhaite remercier Fab. Mon Fab, tu as été mon roc en particulier au cours des difficiles derniers mois de cette thèse. Merci d’avoir toujours été là pour moi dans ces moments compliqués, de m’avoir écouté râler, d’avoir géré toute la logistique pour que je n’aie pas à m’en préoccuper. Et merci pour ta précieuse relecture de mon manuscrit.
Merci aussi à mes parents, qui m’ont toujours soutenue dans mes études, vétérinaires d’abord, puis ce doctorat. Et merci à mon (petit) frère Jean-Eudes, qui entre dans le supérieur en même temps que j’en sors, pour son soutien à base de vannes plus ou moins subtiles.
Merci à mes amies, Barbara, Pauline (N°2), Constance, Marine, Samirah. Barbara, tu m’as soutenue même quand toi-même avais besoin de soutien, et on s’est épaulées au cours des années passées. Merci d’être toujours là ! Pauline (number 2 of course), ma co-galérienne préférée, merci pour ces trois années de délires et de potins du midi, et de décompression le weekend.
Merci à Jean-Noël et Carmen, qui m’ont eux aussi soutenue au cours de l’année passée. Jean-Noël, un immense merci pour tes conseils concernant mon avenir professionnel, et pour les délicieux déjeuners à Montesson qui me permettaient de souffler l’espace d’une petite heure pendant la rédaction.
Enfin, un grand merci à Fanny qui m’a aiguillée vers la recherche, et à son mari Sébastien qui m’a fait découvrir le travail de laboratoire il y a maintenant quelques années.
Abstract
Many decisions involve trading effort for reward. Manipulations of locus coeruleus (LC) activity
have demonstrated its causal implication in effort processing. Indirect measurements of LC
activity reflect physical effort and cognitive control exertion and single unit recordings have
provided evidence for physical effort encoding by LC neurons. Moreover, LC neurons respond
to commitment errors and pupillometry studies and non-specific enhancements of LC activity
showed its role in adjusting effort allocation based on action consequences. We hypothesized
that LC neurons encode physical and mental effort costs, facilitate future effort production,
and participate in adjustments of effort allocation strategies based on previous effort costs
and action consequences. Here, the activity of LC units recorded in 4 tasks was examined. LC
neurons activated with temporal dynamics coherent with responses to the termination of
decisions, and consistently encoded the effort costs of triggering and executing actions. We
confirmed that the timing of these modulations was only compatible with delayed effects on
effort production. Moreover, in a two-alternative choice task, increased LC activity early in
trials was associated to within-trial facilitation of effort production. Additionally, LC neurons
responded to execution errors and increased LC activity around action execution was
associated to increases in effort sensitivity in the next trial. We discuss the behavioral
implications of such effects, their relation with other demonstrated roles of LC activity such as
behavioral variability and set-shifting, and the potential inputs and targets responsible of
them.
1
Résumé
De nombreuses décisions impliquent un compromis entre effort et récompense. Les
manipulations de l’activité du locus coeruleus (LC) ont montré son rôle causal dans la gestion
de l’effort. Les mesures indirectes de l’activité du LC reflètent l’effort physique et le contrôle
cognitif mis en œuvre, et les neurones du LC encodent l’effort physique. De plus, les neurones
du LC répondent aux erreurs d’engagement et les études de pupillométrie ou les modifications
non spécifiques de l’activité du LC ont montré son rôle dans l’ajustement du comportement
en fonction des conséquences des actions. Ici, l’activité de neurones du LC a été examinée
dans 4 tâches. Les neurones s’activaient avec une dynamique compatible avec une réponse à
l’achèvement des décisions, et codaient les coûts de l’effort liés au déclenchement et à
l’exécution des actions. Nous avons confirmé que le moment de ces activations n’était
compatible qu’avec des effets retardés sur la production d’effort. De plus, dans une tâche de
choix force/récompense à deux options, les augmentations d’activité du LC étaient associées
à une facilitation de la production d’effort au sein d’un même essai. Par ailleurs, les neurones
du LC répondaient aux erreurs d’exécution et une activité augmentée au moment de
l’exécution des actions était associée à une augmentation de la sensibilité aux coûts de l’effort
dans l’essai suivant. Nous discutons les implications comportementales de ces effets, leur
relation avec d’autres rôles du LC tels que la variabilité comportementale et l’adaptation aux
changements de règles, et les potentielles afférences et efférences mises en jeu.
2
Table of contents
ABSTRACT 0
RESUME 1
I. INTRODUCTION 6
A. Physical and mental effort: characteristics and implications for decision-making 7 1. Physical effort 7
a) Aversiveness of physical effort: physical effort as a cost and decision variable 8 b) Mental representation of physical effort 9 c) Measuring physical effort 9
(1) Ratings 9 (2) Behavioral proxies 10 (3) Physiological proxies 10
2. Mental effort 12 a) Nature of mental effort: definitions 12 b) Categories of tasks requiring cognitive control 13 c) Sub-processes of cognitive control 15 d) Cognitive control and decision-making 16
(1) Cognitive control and optimization of decisions 16 (2) Cognitive control as a decision variable per se 20 (3) Decisions to allocate control and how much control to allocate: the expected value of control
(EVC) theory 24 e) What is the substrate of cognitive control? What is the resource? 31
(1) Limited intrinsic physiological resource 31 (2) Limited computational resource 36 (3) Motivational resource: exploration-exploitation trade-off and opportunity costs of control 37
f) Behavioral proxies for mental effort 38 (1) Reaction time 38 (2) Performance 42
g) Physiological proxies for mental effort: cardiovascular, metabolic and vegetative measures 43 (1) Cardio-vascular and metabolic proxies 43 (2) Pupil dilation 44 (3) EEG proxies for cognitive control 46
3. Mental and physical effort, evidence for common resources and partially common processing systems
48
B. Role of the locus coeruleus in effort processing: review of the literature 50 1. Anatomy, histology and physiology of the locus coeruleus 50
a) Anatomy and histology 50 b) Electrophysiological properties 52 c) Connectivity to other brain structures 53 d) Noradrenergic receptor distribution and actions in target areas 56
2. Assessing LC activity: proxies for LC activity 57 a) P3 ERP 57 b) Pupil dilation 59
3. LC activity and enhancement of processes requiring resource mobilization 61 a) LC activity and arousal 61
3
b) LC and stress-related responses 63 c) LC activity and attention 64
4. LC and execution of costly goal-directed actions 66 5. LC and decision-making 74
a) Action selection 75 b) Action triggering 81
6. LC activity and action monitoring 87
C. Mechanisms for the effects of LC activity on target areas to regulate cognitive processes, and
implications for effort processing 89 1. LC activity modulates brain metabolism: NA and response to high metabolic demands 89 2. From LC facilitation of sensory processing and plasticity to gain theory 90
a) Facilitation of sensory processing 90 b) Overarching theories of short-term effects of NA on neuronal activity: brain states and gain
modulation 95 c) Extending gain theory in time: LC activity and plasticity promotion 100 d) LC activity and brain connectivity and networks 101
D. Summary and problematic 105
II. EXPERIMENTAL RESULTS 108
A. Study 1: LC neurons encode difficulty to trigger actions in one-option forced choice tasks 108 1. Locus coeruleus neurons encode the subjective difficulty of triggering and executing actions: evidence
for a role in effort processing (Bornert and Bouret, in prep) 108 2. Additional results 153
a) Activity of LC neurons around cue onset and the outcome 153 (1) LC neurons responded to cue onset and the outcome 153 (2) Modulation of LC responses by task parameters at cue onset and the outcome 155 (3) Modulation of LC activity by behavior 157
(a) Cue-evoked activity and willingness to work 157 (b) Real delay and outcome activity 158
b) Dissection of the temporal dynamics of parameter encoding relative to activity changes 158 c) Intrinsic neuronal excitability shaped the activation in response to task events and the encoding of
task parameters and behavioral variables 162 (1) Higher baseline was associated to higher peak of activity 162 (2) Higher baseline was associated to higher signal-to-noise ratio 164 (3) Higher baseline was associated to stronger encoding of task parameters and behavioral variables
166 (4) Encoding of Reward discounted value 167
3. Discussion of additional results and of study limitations 167
B. Study 2: LC neurons facilitate delayed effort mobilization and monitoring in a two-alternative choice
task 171 1. Introduction 171 2. Materials and methods 174
a) Animals 174 b) Behavioral paradigm 174 c) MRI, surgery and recordings 176 d) Data analysis 178
(1) Eye-tracking and determination of action-selection and action-triggering time 178 (2) Value inference from choices 180
3. Results 181
4
a) Behavior 181 (1) Engagement in costly actions 183
(a) RT to the middle press represented the cost of triggering the middle action and starting a trial
183 (b) Decisions to engage based on cue information 185
(i) Choice engagement increased with increased sum of Rewards of options 185 (i) Action-triggering time scaled with difficulty to engage in the choice 187 (ii) Difficulty to engage in the choices scaled with difficulty to engage in the middle press 190
(2) Choosing between offers 193 (a) Influence of Forces and Rewards on binary choices between offers 193 (b) Action-selection time had a tendency to increase with choice difficulty 195 (c) Action-selection time did not influence choice accuracy 197 (d) Absence of relationship between engagement difficulty and choice accuracy 198 (e) Increased action-selection time is associated to decreased sensitivity of choices to Forces 199
(3) Execution of costly actions 202 (a) Middle press exerted force was negatively related to difficulty to engage in middle press 202 (b) Side press exerted force was negatively related to difficulty to engage in side press 204 (c) Performance was negatively related to action-triggering time 204
(4) Summary of behavioral results 205 b) Neurophysiology of the LC 206
(1) Middle press-related activity: encoding of middle press RT and enhancement of engagement-
related processes and of Force exertion 208 (a) LC neurons increased their rate before and during middle press 208 (b) LC neurons encoded middle press RT before middle press onset 209 (c) Middle press activity and engagement throughout the trial: higher pre-middle press activity
was associated to higher probability to engagement in the choice 212 (d) Middle press activity and force exertion: higher pre-middle press activity was associated to
increased middle press exerted force 216 (2) Activity around cues onset 217
(a) Responses to cues onset and choice onset 217 (b) LC neurons positively encoded immediate subjective difficulty to engage in choice around
cues onset 219 (c) LC neurons encoded Rewards but not Forces after cues onset 222 (d) Increased activity of LC neurons around cues onset was associated to increased probability to
choose high-Force option 224 (e) Activity around cues onset had a positive effect on side press exerted force in high-chosen
Force conditions 227 (3) Side-press related LC activity 230
(a) Evoked activity around action onset 230 (b) LC neurons encoded the difficulty of the action during its execution 231 (c) Higher press-related LC activity at trial N was associated to increased Force sensitivity of
choices at trial N+1 234 (4) Inter-trial activity of LC neurons 236
(a) LC activity increased during the ITI following failures to execute the side press 236
III. DISCUSSION 240
A. Behavior: signatures of effort in one- and two-alternative tasks 241 1. Cognitive control exertion for decisions 241 2. Cognitive control for triggering actions 243 3. Physical effort to execute presses 244
5
B. LC activation may signal the end of decision processes 245
C. LC neurons encode subjective difficulty, and possibly effort, to engage in actions and execute them 249
D. LC activity and value-based decision-making 252
E. LC and facilitation of effort exertion: mechanistic hypotheses 257
F. Relationship between noradrenergic facilitation of effort and promotion of behavioral variability 260
G. Modularity of the LC: evidence for differences in electrophysiological properties and consequences for
effort processing 261
H. Conclusions and perspectives 263
REFERENCES 264
LIST OF FIGURES 305
LIST OF TABLES 309
LIST OF ABBREVIATIONS 310
RÉSUMÉ EN FRANÇAIS 311
ANNEX: PHARMACOLOGICAL EVIDENCE FOR THE IMPLICATION OF NORADRENALINE IN EFFORT,
BORDERIES ET AL., 2020 327
6
I. INTRODUCTION
“Success is dependent on effort” – Sophocles
Will you choose to make the effort to read this manuscript, or will you choose to watch
a movie? Every day, several times a day, we face decisions about actions involving effort
exertion (in my example, deciding between reading my manuscript or watching a movie), and,
depending on the result of such decision, effort exertion (if you choose to read it, you will have
to make the effort to read it). In the Cambridge English dictionary, an effort is defined as a
‘physical or mental activity needed to achieve something’. ‘Being an effort’ is defined as being
‘difficult, tiring or boring to do’. In addition, the Oxford dictionary defines effort as a ‘strenuous
physical or mental exertion’. Put together, these definitions emphasize several characteristics
of effort. First, effort is goal-directed, as it aims at achieving an objective. Effort is indeed
usually exerted in order to reach a goal that would not have been reached without effort
exertion, e.g. reading a PhD thesis. Second, effort is aversive (difficult, tiring or boring), and
would be avoided if possible, e.g. you would definitely rather be watching a movie than
reading these words. This introduces that subjects use effort as a decision variable, i.e. effort
influences choices: if the same goal could be achieved through two courses of action, the least
effortful one will be chosen. For example, a student certain that he does not need to study for
a test to score an A grade will probably not study, as it would constitute an effort, while the
goal (scoring an A) could have been achieved without it. Third, making an effort is associated
to spending resources (strenuous) and extensive or intensive effort can lead to fatigue. This
provides a justification for the aversiveness of effort, although the nature of the resource is
not specified by the definitions. Fourth, both definitions draw a line between mental and
physical effort, although this distinction has long been debated. Dewey (1897), posited that
this distinction is merely a distinction based on the nature of the goal (moral or physical) rather
than a psychological one, and that any effort is intellectual.
Effort, in neuroscience, is generally opposed to rewards, i.e. the benefits of actions.
While the brain structures and neurotransmitters involved in reward processing are now well
known and have been extensively studied (see Ikemoto et al., 2015 for a review of the role of
dopamine in reward processing), behavioral and cognitive neurosciences have not yet entirely
solved the question of the neuronal basis of effort, nor of the nature of the resources
7
expensed. This work, however, intends to contribute to this vast subject, by proposing a
candidate neuronal substrate for processing and exerting effort: the locus coeruleus (LC)
noradrenergic brain system. In order to provide the reader with the context of this hypothesis,
the first part of this manuscript will consist in an extensive review, first, of the literature
regarding physical and mental effort, and their common points. It will be followed by a brief
summary of the anatomo-functional properties of the LC, before an extensive review of the
various studies that led to formulating the hypothesis of the implication of the LC in effort
processing. The corpus of the manuscript then includes the reports of two experimental
studies aiming at further exploring the relationship between LC activity and effort processing
and exertion in macaques. I finish with a general discussion summarizing the findings of the
experimental studies, addressing the broader impact of this work, its limitations, and
providing perspectives for future research.
A. Physical and mental effort: characteristics and
implications for decision-making
1. Physical effort
In everyday life as well as in neuroscience research, effort often refers to a physical
requirement. Understandably, most rodent and non-human primate studies on effort
processing have focused on tasks such as force exertion, barrier jumping or repeated lever
presses. This operationalization of effort has been extended to human research, with many
effort studies in humans involving repeated button presses (Porat et al., 2014; Treadway et
al., 2009) or pressing a hand-held dynamometer with variable strength (Bonnelle et al., 2016;
Chong, 2015; Chong et al., 2015; Cléry-Melin et al., 2011; Kurniawan et al., 2010; Prévost et
al., 2010). In this part, I review characteristics of physical effort relevant for decision-making
i.e. its perception as a cost and its use as a decision variable. I then briefly review studies
regarding the mental representation of physical effort. Finally, I provide an overview of the
proxies and physiological measurements than can be used as measurements of physical effort.
8
a) Aversiveness of physical effort: physical effort as a
cost and decision variable
Neuro-economics conceptualize decisions as relying on a cost-benefit trade-off
between the costs and benefits of offers, aiming at maximizing the utility of the chosen option
(Von Neumann and Morgenstern, 2007). Hull (1943) formulated the law of less work, that
states that effort intrinsically carries disutility:
“If two or more behavioral sequences, each involving a different amount of energy
consumption or work, have been equally well reinforced an equal number of times, the
organism will gradually learn to choose the less laborious behavior sequence leading to the
attainment of the reinforcing state of affairs.” Hull, 1944
This law has been widely applied to physical effort. Many experimental studies in
animals and humans have demonstrated the costly nature of effort, by repeatedly showing
that subjects required higher physical effort options to be associated with higher rewards to
choose them (e.g. Cousins and Salamone, 1994; Salamone et al., 2007; Stevens et al., 2005;
Walton et al., 2006). An example of such paradigms and the associated results is presented in
Figure 1.
(A) Walton et al., 2006, Figure 2A (B) Walton et al., 2006, Figure 2B
Figure 1 : Physical effort carries disutility and increased effort requirements need to be
associated with increased rewards to be chosen (adapted from Walton et al., 2006)
(A) Example of task design to demonstrate the disutility of physical effort. Rodents were
offered two alternative courses of action: a high effort-high reward (HR) and a low effort-low
reward option (LR). Both required to press a lever for a certain amount of times (FR, 4 for the
low reward lever, 4 to 20 for the high reward lever) to receive a certain reward (2 pellets for
the low reward lever, 4 for the high effort lever).
(B) Behavioral results: proportion of high reward (HR, y-axis) choices depending on the number
of presses (FR, x-axis) required to get the reward. Increasing press number led to a strong
9
decrease in the proportion of high effort-high-reward choices, indicating that the high reward
was depreciated by the increased physical effort.
b) Mental representation of physical effort
In the past, physical effort perception had been theorized to critically depend on
peripheral sensory afferences from the muscles and organs implicated in the execution of the
effort, but several studies led to the questioning of this idea. Specifically, physical effort
perception was not altered by removing afferent sensory feedback via epidural anesthesia of
afferent nerve severing (Braith et al., 1992; Smith et al., 2003; Zhao et al., 2003; reviewed by
Marcora, 2009). Moreover, it was increased by curare1 administration, which increased the
intensity of the central motor command necessary to maintain constant muscular contraction
(Gallagher et al., 2001). A concurrent hypothesis was proposed: the somatosensory cortex
could directly receive a copy of the motor command (de Morree et al., 2012; Marcora, 2009;
McCloskey, 1981; Proske, 2005). Neuronal disruption studies using transcranial magnetic
stimulation or continuous theta burst stimulation have suggested an implication of the
supplementary motor area (SMA) in effort perception, as disrupting SMA activity led to an
increase in exerted force (White et al., 2013) or a decrease in perceived physical effort (Zénon
et al., 2015).
c) Measuring physical effort
(1) Ratings
Due to the subjective nature of physical effort, many studies have explored potential
proxies for it. The most obvious measurements of subjective physical effort consists in asking
participants to rate their effort exertion (rate of perceived exertion, RPE), and RPEs constitute
a gold-standard measurement (Borg, 1982). Note that the fact that physical effort was shown
not to be perceived via sensory feedback suggests that rating questionnaires should be
formulated with caution so that participants do not rate the physical unpleasantness of the
action (e.g. muscle ache) but the subjective effort.
1 Curares are alkaloids, high affinity antagonists of the nicotinic receptor of acetylcholine. Their fixation on this receptor in the neuromuscular junction blocks the effects of acetylcholine and partially or completely (depending on the dosage) prevent muscular contraction (Manalis, 1977).
10
(2) Behavioral proxies
In the absence of ratings (because the task did not involve ratings, or in the context of
animal research), in tasks offering the possibility to refuse trials, the acceptance rate has been
used to evaluate the effort represented by options (e.g. Bowman et al., 1996; Jahn et al., 2020;
Salamone et al., 2007; Shidara et al., 2005; Varazzani et al., 2015; Walton et al., 2009; Zénon
et al., 2014, Figure 2B).
(A) Zénon et al., 2014, Figure 3D (B) Zénon et al., 2014, Figure 3E
Figure 2 : Acceptance rate decreased when physical effort increased in a force exertion task
consisting in pressing a grip at four different force levels (adapted from Zénon et al., 2014)
(A) RPEs for the 4 conditions across subjects: subjects reported perceiving a greater effort for
greater required force.
(B) Acceptance rates by condition: acceptance rates decreased with increased force.
A comparable strategy consists in evaluating the frequency of choosing an effortful
option when the alternative is a no-effort option (Salamone et al., 2007; Stevens et al., 2005;
Walton et al., 2006). In both cases, those measures reflect the anticipated effort (a priori effort
evaluation) and not the perceived effort (a posteriori effort evaluation), and do not offer trial-
by-trial resolution but rather session-wide assessments by trial category. Alternatively,
another proxy for the subjective effort of options, again available across trials, is the rate of
accepting to replicate a given effort i.e. how often subjects accept to replicate a given effort
for different amounts of reward (Zénon et al., 2014).
(3) Physiological proxies
Pupil diameter was proposed as a proxy for physical effort after accounts of increased
pupil diameter in response to increase mental effort (see part I.A.2.g)(2)). In a force exertion
task, Zénon and colleagues (2014) showed that pupil diameter during force exertion scaled
with the required force (after an initial dip that was identical across conditions, Figure 3). A
similar result was obtained in monkeys performing a force task (pressing a grip to get a reward)
(Varazzani et al., 2015). Moreover, by asking subjects to replicate their presses, the study
11
showed that the intensity of the pupil dilation during the first press negatively predicted the
probability to accept to replicate the press, even after factoring out the effects of the strength
applied during the first press. Coherently, pupil dilation predicted RPEs, even when including
the exerted force in the model. Altogether, the results of this study indicate that pupil
diameter could be a good proxy for the subjective effort of individual trials. Such a relationship
between physical effort and pupil diameter is of particular interest for this manuscript
considering the known relationship between LC activity and pupil diameter (see part I.B.2.b)).
Heart rate, oxygen uptake and blood lactate have also been proposed as a proxies for
subjective physical effort. Several studies have demonstrated a correlation between RPEs and
such peripheral measures reflecting metabolism (Andrade et al., 2020; Da Silva et al., 2020;
Tibana et al., 2019, Figure 4). However, due to slow variations and, for blood lactate
measurement, invasiveness of the measurement (blood sampling), such measures lack the
temporal resolution required to evaluate trial-by-trial physical effort in the context of brief
and relatively low-intensity actions involving few muscular groups (pressing a grip for
instance).
(A) Zénon et al., 2014, Figure 3A (B) Zénon et al., 2014, Figure 3B
(C) Zénon et al., 2014, Figure 3C (D) Zénon et al., 2014, Figure 4B
Figure 3 : During force exertion, after an initial dip, pupil diameter increased and was positively
correlated with the exerted force (figure adapted from Zénon et al., 2014).
(A) Pupil diameter, aligned onto action onset (grip press), by force category. After press onset,
pupil diameter decreased then increased, with a stronger increase for bigger requested forces.
(B) Mean pupil response (z-scored) by condition. The pupil response was bigger for bigger
requested forces (note that the tendency for an increase was absent between the two lowest
forces).
(C) Mean exerted force, aligned on action onset, by force category (by proportion of the
Maximum Voluntary Contraction – MVC).
(D) Relationship between grip force and pupil response, by subject.
12
(A) Andrade et al., 2020 Figure 2A (B) Andrade et al., 2020, Figure 2B
Figure 4: Positive relationship between RPEs and (A) oxygen uptake (percentage of maximum
uptake, VO2max) and (B) heart rate (percentage of maximum heart rate, HRmax) (adapted from
Andrade et al., 2020).
To summarize, physical effort, materialized by actions mobilizing muscular groups, is
perceived as aversive by subjects (animals and humans), and therefore tends to be avoided if
an alternative associated with less physical effort and sufficient reward is offered. Mental
representations of physical and therefore subjective effort perception appear, interestingly,
not to rely on somatosensory information but to critically depend on the SMA. Subjective
physical effort can be measured, or at least approximated, by asking subjects to rate their
exertion (RPEs). At the session level, the acceptance rate, the rate of choice of the high effort
option when a low effort option is available, or the willingness to replicate the effort can
provide some insight on the perceived physical effort of conditions. Regarding physiological
proxies, pupil dilation during the action is frequently used to assess the subject effort
perceived by subjects, although some have shown that metabolic indices such as heart rate,
oxygen consumption and blood lactate also constitute satisfying proxies.
2. Mental effort
a) Nature of mental effort: definitions
Gaillard defined mental effort as ‘energy mobilization in the service of cognitive goals’
(Gaillard, 2001, 1993) and Kool and Botvinick (2018) as the ‘subjective intensification of mental
activity mediating between how well an organism can potentially perform a task and how well
it actually performs on that task’. In other words, subjects perform mental effort to maximize
subjective utility.
Recent accounts of mental effort have grouped its subcategories under the concept of
cognitive control, defined as a set of cognitive processes facilitating goal-directed behavior by
promoting the active use of task-rules (Cohen et al., 2001; Dixon and Christoff, 2012; Posner
(B)
13
and Snyder, 2004 - reedited version of 1975 seminal paper). Shenhav and colleagues (2017)
proposed that ‘cognitive control can be viewed as the force through which cognitive effort is
exerted’.
b) Categories of tasks requiring cognitive control
Posner and Snyder first restrained the definition of cognitive control to overcoming
strong response biases. Such inhibition of response tendencies can indeed be tested using the
Posner task, a task consisting in detecting the position of a target presented after a distractor,
either invalid or valid (Figure 5A). Valid distractors predict the position of the target, while
invalid distractors attract attention to another location. Subjects have a bias towards
indicating a target position that follows the position of the distractor.
Recently, the notion of cognitive control has however been theorized to include a
larger range of cognitive processes crucial for complex executive tasks: information updating,
inhibition of default responses, and mental set-shifting (Miyake et al., 2000). Updating refers
to the ability to monitor upcoming information, and erase no outdated information. A classical
task mobilizing updating capacities is the N-back task (Kirchner, 1958). The task consists in
presenting a series of stimuli, usually letters, to the participants, who have to indicate whether
the current stimulus is identical to the one presented N trials back (Figure 5B). The N-back task
thus involves trial-by-trial updating of the information stored in working memory, which is a
form of mental effort.
Inhibition is the capacity to suppress response biases when required, i.e. default,
automatic or dominant responses. Such ability can be tested in the aforementioned Posner
task (Figure 5A), or using the Stroop task (Stroop, 1935), a cognitive interference task. Subjects
are required for example to give the color of a written word. The information provided by the
word itself can be congruent (e.g. the word green written in green), and facilitate the color-
naming, or incongruent (e.g. the word green written in blue), and render the color naming
more difficult (Figure 5C). In incongruent trials, both the Posner and Stroop tasks require
ignoring the bias towards responding according to the distractor, thus involving inhibition, a
subcategory of cognitive control.
(A) Posner task: Hayward and Ristic, 2013, Figure 1
(B) N-Back task: Lamichhane et al., 2020, Figure 1
14
(C) Stroop task: Nathan et al., 2012, Figure 2b
(D) Set-shifting task: Monsell et al., 2003, Figure 5
Figure 5: Tasks used to test the different components of cognitive control
(A) The Posner task (figure adapted from Hayward and Ristic, 2013). The objective of this task
is to indicate the position of a target (here, a cross) after its apparition. Importantly, just before
the target appears, a cue attracting attention towards one of the two positions is presented
(here, the bold square). The cue can either be valid (indicate the future position of the target)
or invalid (indicate the other position).
(B) The N-Back task (figure adapted from Lamichhane et al., 2020). The N-back consists stimuli
being presented in a row (e.g. letters) and subjects have to report whether the current one is
identical to the one presented N trials back. The task becomes more difficult for higher N.
(C) The Stroop task (figure adapted from Nathan et al., 2012). The color-word Stroop task
presented here consists in presenting a series of color words (blue, green etc) written in various
colors. The subjects have to report the color of the written word. The written word can be
congruent (green written in green, subjects have to respond ‘green’) or incongruent as
presented in the figure (green written in red, subjects have to respond ‘red’).
(D) Set-shifting task (figure adapted from Monsell et al., 2003). Subjects are presented with a
series of numbers from 1 to 10, displayed on a square. The square can be in two positions (left
side of the figure vs right side of the figure). If the square is placed like the square displayed on
the left of the figure, the subject has to tell whether the number presented is smaller than five.
If the square is in the position presented on the right side of the figure, the subject has to
indicate whether the number presented is even. RT: reaction time.
Finally, switching (set-shifting) corresponds to the capacity to disengage from the task
at hand or from the current task rules to commit to a new task set or use new rules (switch
task and/or rules, Monsell, 1996; Monsell et al., 2003). Set-shifting tasks typically involve
responding to stimuli according to two possible rules. The rule that has to be used for each
stimulus is indicated by a cue. For example, Monsell and colleagues (2003) used a task
consisting in presenting a series of numbers from one to ten to the subjects. Numbers could
be presented superimposed onto two different shapes, indicating the rule that had to be
followed to answer: subjects either had to tell if the number was even or if it was smaller than
15
five (Figure 5D). Switching between task rules requires engaging cognitive control. It is
important to note that task-switching is quite dependent on inhibitory control in a large
variety of tasks. Indeed, inhibitory processes are required to deactivate the no longer needed
task rules, a process known as backwards inhibition (Dajani and Uddin, 2015; Klimesch, 2011;
Mayr and Keele, 2000; Zink et al., 2019).
To summarize, tasks requiring cognitive control exertion can generally be put into
three possible sub-categories: updating (working memory) tasks, in which control is used to
renew the information stored in working memory, inhibition tasks, in which control is required
to block default responses, and switching tasks, in which task rules change at a fast pace and
control is required to adjust.
c) Sub-processes of cognitive control
Analyzing the characteristics of the tasks requiring cognitive control exertion allowed
Shenhav and colleagues (2013) to dissect it and identify three core component functions:
regulation, specification and monitoring. Regulation refers to the capacity of the control
system to “influence lower-level information-processing mechanisms” by adjusting the identity
of the attended information, the target parameter for the regulation, and the intensity with
which the processing system attends to that information, i.e. the degree of displacement of
the parameter. For example, in a Stroop task, regulation is involved in attending to the correct
stimulus (in case of a color task, the color of the word and not the word itself) and modulating
the strength with which it is inspected. Beyond this specific example, the regulation
component can modify various other parameters. For instance, it can modulate the thresholds
and biases implemented for decision-making (Bogacz et al., 2006; Wiecki and Frank, 2013),
the searching patterns implemented for searching through visual stimuli or memory
information (Olivers et al., 2011; Polyn et al., 2009), or regulate negative emotions (Gross,
2011; Johns et al., 2008).
Specification refers to decisions about engagement in control demanding tasks, about
goals that should be pursued and tasks that should be executed, and about control allocation.
Consequently, this process specifies identity and intensity for the regulation component.
Specification is distinct and upstream from regulation since specification determines the
identity and intensity of the control signal while regulation merely implements them. Details
16
regarding decisions to allocate control and their implementation will be provided in part
I.A.2.d)(3).
Monitoring happens downstream from regulation, as it serves to process information
about the result of the control exertion on task execution in order to optimize future
specification, using feedbacks such as reaction time (RT), reward, pain, delay to reward, errors,
or other negative feedbacks. Note that the monitoring unit does not provide the feedback
about the control exertion, but responds to that feedback. For instance, Botvinick and
colleagues (Botvinick, 2007; Botvinick et al., 2004, 2001) focused on cases in which a response
conflict is detected (possibly by monitoring RT), which can be considered as signaling
insufficient control exertion. In such situations, the monitoring system needs to detect the
conflict and adjust the parameters of the specification system, in a form of reinforcement
learning. In that frame, allocating control can be reframed as an optimization problem, with
the objective of maximizing short- and long-term utility (Bogacz et al., 2006; Dayan, 2012; Hazy
et al., 2007; O’Reilly and Frank, 2006; Todd et al., 2008; Yu et al., 2009).
To summarize, the cognitive control system has been hypothesized to be constituted
of three subunits: a specification unit deciding on the identity and intensity of the control, a
regulation unit implementing them, and a monitoring unit receiving feedback on the results
of the control exertion in order to adjust the decisions made by the specification unit.
d) Cognitive control and decision-making
(1) Cognitive control and optimization of decisions
Numerous behavioral studies in healthy patients (e.g. Hinson et al., 2003; Missier et
al., 2012, 2010), patients with frontal cortex lesions (e.g. Eslinger and Damasio, 1985; Manes
et al., 2002) or with attention-deficit hyperactivity disorder (ADHD, e.g. Mäntylä et al., 2012),
as well as brain imaging studies (e.g. Clark et al., 2004; Coricelli et al., 2005; De Martino et al.,
2006; De Neys et al., 2008) have pointed towards a link between various subcategories of
decision-making and cognitive control capacity. In addition, pupil dilation, a proxy for mental
effort (see part I.A.2.g)(2) for review), has been shown to increase during decisions (Beatty,
1982; de Gee et al., 2014; Gilzenrat et al., 2010; Lempert et al., 2015; Nassar et al., 2012; Figure
6A).
17
Removing any possible confound with the motor preparation of the response,
Einhäuser and colleagues (2010) asked subjects to make covert choices between cues (digits)
sequentially presented on screen, and to report the chosen digit only at the end of the trial.
Although the decision to choose a digit when it was presented did not involve immediately
producing a movement to report the choice, pupil diameter specifically increased during
presentation of the digit later reported as chosen. Pupil diameter has also been reported to
increase around the time of the end of decision processes, materialized by a motor action to
respond, in tasks consisting in continuously tracking perceptual stimuli and reporting
detecting changes (Einhäuser et al., 2008; Hupé et al., 2009). Interestingly, pupil diameter has
been shown to be larger during decisions in which subjects end up responding against their
bias (de Gee et al., 2020, 2017, 2014) or have a low confidence, i.e. difficult choices (Lempert
et al., 2015). Altogether, these results suggest that cognitive control exertion, approximated
by pupil dilation, can improve decision-making, particularly by facilitating responses away
from the default (responding according to existing biases).
(A) Lempert et al., 2015, Figure 3A (B) De Gee et al., 2017, Figure 2A middle pannel
Figure 6: Pupil diameter increases during decisions, especially when subjects will respond
against their bias.
(A) Time course of pupil dilation, aligned on stimulus onset, in an auditory decision task. Pupil
increased after stimulus onset, during the decision period. Dashed lines represent the standard
error of the mean (SEM), and dashed line represents mean across subjects (adapted from
Lempert et al., 2015).
(B) Quantification of response biases in trials with small task evoked pupillary responses (TPR,
blue) and big TPR (red) (median split), in a visual detection task. Increased task-evoked
pupillary responses (TPR) were associated with decreased response biases (adapted from De
Gee, 2017).
The predominant theory regarding the control mechanism applied to decision-making
is the ‘dual process theory’ (e.g. Epstein, 1994; Evans, 1984; Evans and Over, 2013; Evans et
al., 2007; Goel, 1995; Kahneman and Frederick, 2005; Sloman, 1996; Stanovich and West,
2000). This theoretical framework intends to explain failures to think rationally observed in
18
human behavior. De Neys and Glumicic (2008) provide an insightful example of such failure
by citing the case of a physical aggression in a European country being attributed to an ethnical
minority (African immigrants) during a period of high racial tensions. Although the small
relative proportion of African immigrants in the general population made the probability of
the aggression being perpetrated by one of them highly unlikely, the African community was
immediately blamed due to irrelevant priors regarding that community. The example is
particularly striking considering that the aggressors were later found to be Caucasian
Europeans.
One general control mechanism for decision-making has been proposed, with small
disagreements over details among authors. Dual process theory states that two systems using
different sources of information and operating at different paces compete to formulate a
choice. First, the heuristic system, relying on prior beliefs and knowledge, operates fast and
automatically, as its operations are not very computationally demanding. Conversely, the
analytic system uses demanding computations, is slower, and allows rational and logical
reasoning. In the previous example, the heuristic system pointed towards accusing African
immigrants, while the analytic system would have indicated that this reasoning was illogical.
In that specific example, a conflict arises between the two systems, and cognitive control is
believed to be required for the analytic system, using costly computations, to override the
belief-based response of the heuristic system.
There is still an open debate on whether the exertion of control allows merely detecting
the conflict i.e. being aware of one’s biases and prejudices and heuristic computations (Evans,
1984; Kahneman and Frederick, 2005), or overriding the default in case of a conflict, i.e.
favoring analytic computations (Epstein, 1994; Shenhav et al., 2013; Sloman, 1996). The latter
intuition relies on post-decisional reports from participants that they ‘knew’ their answer was
wrong but chose it because it ‘felt’ right (Denes-Raj and Epstein, 1994; Epstein, 1994; Epstein
and Pacini, 1999), suggesting that being aware of one’s priors is not costly, but overriding them
is. Some have intended to solve that debate by asking subjects to ‘think aloud’ while making
a decision (De Neys and Glumicic, 2008). Participants were asked to answer a question by
choosing between two alternative options, one answer being likely to be chosen if they
predominantly used the heuristic system and the other if they predominantly used the analytic
system. For example:
19
“In a study 1000 people were tested. Among the participants there were 5 men and 995
women. Jo is a randomly chosen participant of this study.
Jo is 23 years old and is finishing a degree in engineering. On Friday nights, Jo likes to go out
cruising with friends while listening to loud music and drinking beer.
What is most likely?
a. Jo is a man
b. Jo is a woman” (De Neys and Glumicic, 2008)
As expected from classical findings regarding sub-optimality of decision-making and
susceptibility of decision-makers to priors, participants, in most cases, picked the answer that
followed their prior regarding gender-related behavior, thus ignoring the low probability of
‘Jo’ being male, due to the large predominance of women in the described group. The thinking
aloud procedure however provided further insight on the interplay between the heuristic and
analytic systems, as authors report participants phrasing their reasoning as, e.g.: “ … because
Jo’s drinking beer and loud I guess Jo’ll be a guy, although there were more women…”.
Participants thus referred to the relevant information, and authors argue that they were
therefore aware of the result of the analytic computation but choose to ignore it, because too
much cognitive control would be needed to override the default answer from the heuristic
system. Consequently, their results are in favor of control serving to override heuristic
computations, not detecting a conflict between the two systems.
Additionally, comparison between the decision-making tasks highly dependent on
cognitive control (e.g. Dretsch and Tipples, 2008; Hinson et al., 2003; Missier et al., 2010; Shiv
and Fedorikhin, 1999) and tasks requiring low levels of cognitive control exertion (e.g.
Glöckner and Witteman, 2010; Hogarth, 2005; Stanovich and West, 2000) for good
performance has allowed identifying features of both categories (Botvinick and Cohen, 2014;
Cohen et al., 1990; Missier et al., 2012). Shiffrin and Schneider (1977) theorized that
information processing can be more or less automatized, and this continuum of automaticity
in task solving could suffice to categorize tasks along the high to low cognitive control demand
tasks axis. For instance, tasks requiring high levels of cognitive control fall under the category
of tasks requiring extensive processing of decision options, and consequences of choices.
Conversely, tasks requiring low amounts of cognitive control can be solved using low-level
20
computations, e.g. perceptual cues, and automatic processing, e.g. heavily practiced tasks,
learned associations, and recognition processes.
To summarize, decision-making is frequently associated with cognitive control
exertion, especially in the context of decisions requiring extensive information processing, as
shown by studies using pupil dilation as a proxy. Specifically, control is (probably) exerted to
override default responses emerging from low-level computations (heuristic system) and
decide based on more costly and complex computations (analytic system), at least when a
conflict arises between those two alternative responses.
(2) Cognitive control as a decision variable per se
“We like to solve problems easily. We can do so best if we can fit them rapidly into a
satisfactory category and use this category as a means of prejudging the solution.... So long as
we can get away with coarse overgeneralizations we tend to do so. Why? Well, it takes less
effort, and effort, except in the area of our most intense interests, is disagreeable.”
Allport, 1954
Decision-making involves weighing the benefits of the options offered against their
predicted costs (Kahneman and Tversky, 1979; Stephens and Krebs, 1986). Cognitive control
acts as a decision variable, as the “law of least effort” (Hull, 1943) applying to physical effort
applies to mental effort as well (Zipf, 1949). We saw in the previous part (I.A.2.d.1) that
applying cognitive control is beneficial for decision-making processes, especially complex
ones, requiring integration of important amounts of information to produce an accurate
response. However, in that context, humans tend to rely on simplifying strategies of
information gathering and processing (Gigerenzer, 2010; Gigerenzer and Goldstein, 1996;
Sidarus et al., 2019; Simon, 1955; Tversky and Kahneman, 1974). These heuristics (mental
shortcuts) are believed to serve the purpose of reducing the cost of control that would be
associated with the alternative optimal but computationally demanding decision strategy. In
that sense, decision-makers compute a trade-off between the cost of control exertion and the
optimality of their decisions. In that part, I will show how this finding can be generalized
beyond decision-making tasks, and review evidence that applying control is aversive and
serves as a decision variable.
21
Applying cognitive control is aversive (Kool et al., 2010; Kool and Botvinick, 2018,
2014), and courses of action associated with higher cognitive control demands are less likely
to be chosen. For example, Kool and colleagues (2010) designed a choice task to test for
cognitive demand avoidance. The task consisted in repeatedly choosing between pairs of cues,
without previous instructions on their characteristics. Choosing either cue revealed a number
between one and nine. The color of its display indicated if the subject had to tell whether the
number was greater than five (blue number, ‘yes’ or ‘no’) or whether the number was even
(purple number, ‘yes’ or ‘no’, Figure 7A). Importantly, one cue led to a low probability of task
switching, i.e. low-demand cue, and the other cue led to a high probability of switching
assignment, i.e. high-demand cue. Authors report that, throughout a session, subjects learned
to choose the low-demand cue more often (Figure 7B), indicating that subjects perceived the
high-demand cue as carrying an intrinsic negative utility. In this example, mental effort was
probably not required for the choice between cues itself, nor did it influence its quality, as the
cognitive effort was applied as a consequence of the decision, depending on the difficulty of
the given assignment. The rule of least effort seems to have applied here, and to be
extendable to the specific context of decisions requiring investing cognitive control.
Further proof of the negative utility of cognitive control exertion stems from the
observation that increased incentive (offered reward or intrinsic motivation) can compensate
the aversiveness of effort, as increased incentive in cognitive effort tasks leads to increased
performance, within subjects. Such an effect was demonstrated in various types of tasks
relying on the different components of cognitive control:
- updating (working memory tasks, Beck et al., 2010; Heitz et al., 2008; Jimura et al.,
2010; Kennerley and Wallis, 2009; Libby and Lipe, 1992; Taylor et al., 2004; Westbrook et al.,
2013),
- inhibition (Stroop task: Krebs et al., 2010, Posner task: Engelmann et al., 2009;
Modified Stroop task: Padmala and Pessoa, 2011, see Figure 8A, Dixon and Christoff, 2012,
Stop signal response task: Boehler et al., 2014; Leotti and Wager, 2010; Padmala and Pessoa,
2010),
- set-shifting (Aarts et al., 2010; Dixon and Christoff, 2012; Kleinsorge and Rinkenauer,
2012; Nieuwenhuis and Monsell, 2002; Umemoto and Holroyd, 2015). More generally, it was
demonstrated that intelligence test scores could be influenced by the motivational state of
the subjects (Duckworth, 2011).
22
(A) Kool and Botvinick, 2018, Figure 1A (B) Kool and Botvinick, Figure 1B
Figure 7: Behavioral demonstration of the aversiveness of mental effort (figure adapted from
Kool and Botvinick, 2018)
(A) Task design. Subjects were repetitively asked to choose between pairs of previously
unknown cues. When a cue was chosen, it revealed a number from 1 to 9 written either in blue
(rule: ‘state whether the number is greater than 5’) or purple (rule: ‘state whether the number
is even’). Each pair of cues (8 pairs throughout task) was associated to either a high (high-
demand cue) or low (low-demand cue) probability of task (and therefore number color)
switching.
(B) Behavioral results: throughout the task (x-axis, trial number), subjects learned to pick
preferentially the cue associated to a low cognitive demand (y-axis, probability of choosing the
low-demand cue).
Note that these enhancements of performance in response to increased incentive
were not observed in every situation, as a few conditions had to be fulfilled. First, adjusting
the intensity and direction of the control exerted had to be beneficial for task performance
(Camerer and Hogarth, 1999). For instance, task difficulty had to be compatible with the
individual skills and capacities of the subjects (Awasthi and Pratt, 1990; Bonner et al., 2000;
Rydval and Ortmann, 2004, 2004). Moreover, excessively big or small reward size could be
detrimental to performance, as small rewards could seem insulting (Gneezy and Rustichini,
2000) and huge rewards could make subjects ‘choke under pressure’ (Ariely et al., 2009;
Baumeister, 1984; Beilock, 2010; Bonner et al., 2000; Camerer and Hogarth, 1999; Mobbs et
al., 2009; Samuels and Whitecotton, 2011; Worthy et al., 2009). Nonetheless, beyond these
relatively anecdotal exceptions, the general rule that rewards incentivize control allocation
provides further evidence for the negative utility of control.
(A) Shenhav et al. 2017, Figure 2D top panel
(B) Shenhav et al. 2017, Figure 2D bottom left panel
(C) Shenhav et al. 2017, Figure 2D bottom right pannel
23
Figure 8 : Results of the modified Stroop task used by Padmala and Pessoa (2011). RTs were
shorter, accuracy was higher and difficulty had lesser effects on RT and accuracy when subjects
were rewarded.
(A) Selective attention task. Subjects were first informed by a cue of the potential reward of
the following trial. They were then presented with images either of houses or of buildings. A
task irrelevant word was written on top of the image, and could either be incongruent (house
written on a building image), congruent (house written on a house image) or neutral (random
word). Subjects had to tell whether the image represented a house or building. They were told
to respond as quickly and accurately as they could. NB: the numbers presented on the fifth
frame of the task are a feedback on the number of points earned in the trial (top) and since the
beginning of the task (bottom)
(B) RTs of subjects by difficulty (congruency of the irrelevant word) and reward category
(rewarded or not). Subjects were overall faster in rewarded trials and interference effects were
smaller.
(C) Accuracy of subjects by difficulty (congruency of the irrelevant word) and reward category
(rewarded or not). Subjects were overall more accurate in rewarded trials and interference
effects are smaller.
As a consequence, effort allocation can be analyzed in a neuro-economics frame
(Westbrook and Braver, 2015, 2013). For example, effort-discounting tasks, in which subjects
choose to execute different levels of difficulty of cognitive tasks to obtain different amounts
of offered reward, can allow determining individual indifference points, i.e. the subjective
value of each level of mental effort. Such studies were conducted both with healthy subjects
(Dunn et al., 2016; Kool et al., 2010; Schouppe et al., 2014; Westbrook et al., 2013) and with
psychiatric patients (schizophrenic patients: Culbreth et al., 2016). Note that behavioral
analyses in these tasks have allowed ruling out an interpretation of preferences for low effort
in terms of error (risk) avoidance. An illustration of an example task used to determine the
indifference point for an N-Back task (Westbrook et al., 2013) is provided in Figure 9.
Altogether, these results provide further proof that humans routinely choose not to
invest as much cognitive control as they potentially could, but rather compute a cognitive
control-reward trade-off to decide on the amount of control that is worth investing. In the
specific context of applying cognitive control to make accurate decisions, subjects make a
24
trade-off between cognitive-control related costs and decision accuracy-related benefits. In
that frame, decisions using simplifying heuristics, therefore risking inaccuracy, are not
necessarily suboptimal, as they putatively, in certain environments, correspond to the best
trade-off between expensed control resources and expected rewards (Gigerenzer, 2010; Todd
and Gigerenzer, 2012). This theoretical frame, contradicting the postulate that the best
strategy is the one allowing for maximum possible accuracy, is called ‘bounded rationality’
(Gigerenzer, 2008; Simon, 1997, 1956; Todd and Gigerenzer, 2012). As its name indicates, it
assumes that human rationality is inherently limited by the availability of the resources that
are required to make rational decisions. One major assumption for theory is therefore that
applying cognitive control to any process consume certain resources (see part I.A.2.e) for
discussion on the theories regarding the nature of such resources).
(A) Westbrook et al., 2013, Figure 1, top left panel
(B) Westbrook et al., 2013, Figure 1, bottom left pannel
Figure 9: Cognitive effort titration task (adapted from Westbrook et al., 2013).
(A) On-screen task display. Subjects had to choose between two possible offers involving
different N-back task difficulty levels for different amounts of money.
(B) Depending on the decision made at each trial, a new proposal was made to the subjects, in
order to progressively obtain a more precise evaluation of the subject’s indifference point
(titration). The used algorithm is shown here.
To summarize, cognitive control exertion is aversive, and the theory of bounded
rationality states that subjects compute a cost-benefit trade-off for allocating control.
Consequently, increased rewards can increase control allocation. This strategy does not allow
systematically maximizing rewards, but instead maximizes utility.
(3) Decisions to allocate control and how much control to
allocate: the expected value of control (EVC) theory
I established earlier that exerting control represents a cost, and that decision-makers
compute a trade-off between the expected rewards and costs from potential strategies and
25
intensities of control exertion. The following paragraph will thus focus on the core principles
of the most recent models of mental effort allocation.
Early work on decision-making strategies (Beach and Mitchell, 1978; Payne et al., 1988)
has demonstrated that people trade decision accuracy (probability to make the correct choice
i.e. prospective rewards) for mental effort, i.e. they compute a cost-benefit analysis. Smith
and Walker (1993) have summarized such accounts under the ‘labor theory of cognition’,
adapted from the classical economics ‘labor theory’2. More recently, several teams have used
hierarchical reinforcement-learning theories and artificial intelligence research to formalize
and generalize this theory. They also proposed neuronal bases for their implementation,
which I will not detail here outside of the context of the role of the LC (Dayan, 2012; Frank and
Badre, 2012; Holroyd and McClure, 2015; O’Reilly and Frank, 2006; Shenhav et al., 2017, 2013;
Verguts et al., 2015).
The notion of rational meta-reasoning or resource-rational analysis, i.e. methods to
decide to allocate computational resources, comes from research aiming at developing
intelligent systems interacting with their environment. Such systems have to make the best of
their finite resources (Hay et al., 2014) and have therefore been engineered to maximize the
value of their computations (VOC), i.e. the added utility obtained when more computational
resources are allocated minus the computational resources spent (e.g. Random Access
Memory). This concept was later extended to cognition and brain computations (Griffiths et
al., 2015; Lieder et al., 2014, 2012; Lieder and Griffiths, 2017). Lieder and colleagues have
proposed, and experimentally shown, that individuals learn to predict the VOC of individual
cognitive operations, and select computational strategies (sequences of operations) in order
to maximize the VOC (Figure 10A&B). Such accounts, however, consider operations as more
costly if they merely take longer (i.e. constitute an opportunity cost), while several studies
have shown that subjects can experience different levels of mental effort even when the
duration of trials does not differ (Dixon and Christoff, 2012; Kool et al., 2010; Westbrook et
al., 2013).
2 Labor theory states that the economic value of a good or a service is the consequence of the amount of labor that was required to produce it (theory by Smith, Ricardo and Marx, summarized in Cleveland and Ayres, 2004).
26
The EVC theory complements the VOC theory by introducing a cognitive control
intensity parameter. Practically, as mentioned before (see part I.A.2.c), Shenhav and
colleagues (2017, 2013) proposed that signals implemented by the regulation unit of control
have two dimensions: an identity (what control signal should be selected) and an intensity
(vigor of the control signal). The specification unit of the control system decides upon the
identity and intensity of the control signal based on the expected costs and outcomes, and the
regulation unit implements it. After control execution, the monitoring unit receives feedback
on the consequences of the process (e.g. expensed resources, payoffs) and adjusts the
parameters used by the specification unit to make future decisions about control allocation.
For example, when practicing a Stroop task (giving the name of a picture rather than reading
the word written on top of the picture, Figure 10C), the identity of the control corresponds to
what the subjects is attending to (the picture) while the intensity is the strength with which
the subject attends to the stimulus (Figure 10D). Note that one of the crucial assumptions of
that theory is that increasing the intensity of the control leads to increased costs, thus
justifying that the system cannot invest maximal control to get maximal reward, but has to
compromise. Under that assumption, subjects should select both the strategy and control
intensity that have the highest utility (reward minus control costs, including, but not limited
to, time costs).
(A) Lieder et al., 2014, Figure 2 (B) Shenhav et al., 2017, Figure 2a
(C) Shenhav et al., 2017, Figure 2d top panel (D) Shenhav et al., 2017, Figure 2b
Figure 10 : The VOC model and the EVC model
(A) Lieder and colleagues designed a task in which subjects had to choose between two
strategies (swapping numbers or creating sublists) to sort lists of numbers of different lengths
and different ‘sortedness’ (number of misplaced numbers in list), after training on both
strategies (adapted from Lieder et al., 2014)
(B) The VOC model, in the list sorting task, states that subjects select their sorting strategy by
evaluating, based on the length and ‘sortedness’ of the assigned list, the payoff minus cost
(spent time) of each strategy (adapted from Shenhav et al., 2017).
(C) Task used for the example: picture-naming Stroop task (adapted from Shenhav et al., 2017).
Subjects have to name the picture presented on screen (house or building) and ignore the word
27
printed on top (congruent, incongruent or neutral). NB: 2000 and 42000 are the points earned
by the subject in the current trial and throughout the session.
(D) When doing a picture-naming Stroop task, the identity of the control can be reading the
word, which leads frequently to a negative feedback (due to incongruent trials), or naming the
picture, which often leads to positive feedback. Complementarily, when attending to the name
of the picture, subjects can apply different amounts of control (intensity). Increased control will
lead to increased probability of positive feedback (adapted from Shenhav et al., 2017).
The EVC was mathematically formalized as follows:
EVC (signal, state) = [Σi Pr(outcomei|signal, state).Value(outcomei)] – Cost(signal)
The EVC is a function of two distinct variables, state (or current state, the current situation,
i.e. task difficulty, motivation, etc.) and signal (intensity and identity). Pr is the anticipated
(evaluated) probability of a particular outcome (or future state) following control exertion, in
the current state and when applying the specific control signal. The EVC is the expected value
of the outcome (sum of values of potential outcomes weighed by their probabilities) minus
the cost of the control applied to get these potential outcomes with such probabilities.
Note that the EVC does not only take into account the immediate rewards but also
more distant expected rewards. This can be formalized as follows, in a recursive fashion:
Value(outcome) = ImmediateReward(outcome) + γ.maxi[EVC(signali,outcome)]
The value of the outcome is the sum of the immediate rewards of the outcome and the
discounted (γ, ranging from 0 to 1) maximum expected value of the control that can be applied
in the future states over all feasible signals.
To specify the control signal, the control system must identify the signal identity and
intensity that maximize the EVC (Signal*):
Signal* ← maxi[EVC(signali, state)]
28
The regulation unit then implements this signal maximizing the value of the control. The
monitoring unit is responsible of assessing, through feedback from actual outcomes and
resource expenditure, that Signal* is indeed optimal, and, in the case of sub-optimality,
changes in payoffs or in difficulty, shall order adjustments from the specification system
(Figure 11).
(A) Shenhav et al., 2013, Figure 4A (B) Shenhav et al., 2013, Figure 4B
Figure 11: Representation of the determination of the optimal control intensity signal in a
certain state (payoffs and costs) and in case of increases in payoff or in difficulty (adapted from
Shenhav et al., 2013).
The mathematical equation of the EVC states that the EVC (blue) is the expected payoff of the
control (green) minus the expected cost of the control (red). For a given situation (given payoff
and costs, dashed lined), the specification system determines the intensity of the control that
maximizes the EVC (dashed arrow). Note the EVC can become negative for high intensity
control, i.e. exerting maximum control may carry negative value even though it maximizes the
payoff. In extension, in some situations (very difficult or very low payoff tasks), exerting any
control may carry less utility than exerting no control at all (for example refusing the task).
(A) In case of increase of the payoff (green arrow, full lines), the optimal control intensity
increases (left side shift, full black arrow).
(B) In case of an increase of the difficulty, the probability to get the reward in case of identical
control exertion decreases, i.e. the payoff function is lower, and the control intensity
maximizing the EVC increases (right side shift, full black arrow).
Many other teams have proposed their own version of hierarchical reinforcement
learning models for control allocation, but the core principles remain the same. However,
although the model from Shenhav and colleagues (2017, 2013) is quite exhaustive, it does not
formulate hypotheses regarding the shape of the cost function. This gap was breached by the
work of Kool and Botvinick (2014) that extended economic labor-leisure tradeoff theories
(Nicholson and Snyder, 2012) to mental effort exertion. Labor-leisure models were initially
designed to predict how workers would distribute their time between working hours and
29
leisure hours based on offered wages. Research on labor-leisure allocation before and after
manipulating wages (money earned per hour), while compensating for the difference with a
fixed income (fixed salary plus wage multiplied by number of hours) showed that utility curves
have a concave shape (see Figure 12 for details).
Kool and Botvinick (2014) created a paradigm for studying the mental labor-mental
leisure tradeoff in a comparable fashion. Throughout the testing session, subjects were
offered a choice between a high-demand N-back memory task, requiring exerting cognitive
control, and a low-demand reaction task (responding to specific images with a button press).
Participants performed the task during two separate sessions. The high-demand task was paid
in candy, with different payment systems (base salary and wage per trial, paid in candy)
between sessions. The low-demand task was not paid. The behavior of the subjects after
income-compensated increases or decreases in wage (reward per completed trial of the high-
demand task) confirmed the concave shape of the utility function, suggesting that the cost
function for cognitive control may be following such concave shape.
(A) Kool and Botvinick, 2014, Figure 1A
(B) Kool and Botvinick, 2014, Figure 1B
(C) Kool and Botvinick, 2014, Figure 1C
(D) Kool and Botvinick, 2014, Figure 2A
(E) Kool and Botvinick, 2014, Figure 2B
Figure 12: Labor-leisure theory, in economics and applied to cognitive control (figure adapted
from Kool and Botvinick, 2014).
(A,B,C) Labor-leisure theory in economics.
(A) Concave utility surface representing the utility of each individual labor-leisure combination.
Each black line represents iso-utility. Based on the available hours during which the worker can
work per period (T), and the hourly wage offered (w) a budget constraint is defined (plane in
top plot, projected as a line on bottom plot of figure A). The maximum income is therefore wT
(in case the worker works for T hours). To decide upon his working time, labor-leisure theory
states that the worker will identify the income-leisure combination that carries the highest
utility i.e. intersects the iso-utility line corresponding with the highest utility (red dot).
30
(B) Income-compensated wage decrease: the hourly wage is decreased, and a fixed income is
given so that for an equal number of working hours the worker will earn as much money.
Relatively counter-intuitively, workers tend to decide to work for less hours. The concave shape
of the utility functions explains this finding. Before the income-compensated wage decrease
(blue line), the income-leisure combination carrying the highest utility included more work than
after the decrease (red line, flatter than blue line) since the intersection with the highest iso-
utility line is shifted to the right.
(C) Income-compensated wage increase. The hourly wage is increased and the fixed income is
decreased (green line) compared to the initial payment system (blue line). The point that
maximizes utility after the increase is shifted to the left side, with the worker choosing to work
longer, compared to the baseline, to maximize utility.
(D, E) Labor-leisure theory applied to mental effort. Results of an experiment consisting in
offering the choice between a demanding cognitive task (3-back task with images) and a low
demand reaction task (responding to images of children, signaled by a tone). Subjects were
paid for the 3-back task only, with candy, and the income system was manipulated between
the two sessions of the task.
(D) In session 1 (blue line), the participants received a fixed wage per trial of the high-demand
task. In session 2 (red line), an income compensated wage reduction was imposed (free candy
in addition to small wage per trial). The income-leisure decisions of participants are
represented by the blue and red dots for the high wage and low wage with compensatory
income situations, respectively. In the latter, participants chose to perform less trials and earn
less candy than in the former, although performing the exact same number of trials would
have earned them the same exact amount of candy.
(E) In session 1, participants received a low wage and a high fixed income (blue line). In session
2, the wage was increased and the fixed income was decreased (green line). In the high wage
condition, participants chose to perform more trials of the demanding task compared to the
low wage condition, even though doing the same number of trials in both would have resulted
in equal candy amounts.
In summary, control is allocated rationally by subjects, who perform a cost-benefit
analysis to determine the optimal control strategy (identity) and strength (intensity), weighing
the cost of such allocation and the expected outcomes, should such control be exerted.
31
Shenhav and colleagues have mathematically formalized this tradeoff under the EVC model.
Additionally, generalization of economical labor-leisure theory to mental effort has allowed
demonstrating the concave shape of the cost function of control exertion.
e) What is the substrate of cognitive control? What is
the resource?
Overall, the extensive body of literature detailed above provides strong evidence for
the aversive nature of cognitive control exertion, and demonstrates that subjects make
decisions to allocate control as if sparing a depletable resource. One question however
remains, raised by the very formulation of the theory of bounded rationality. If human
cognitive control exertion is bounded by the availability of certain resources, what is the
nature of such limited resources? Two competing yet not incompatible theories have been
formulated, which I will both review. The first posits that cognitive control uses limited
intrinsic resources (metabolic or computational, synthetic representation of these accounts in
Figure 13). The second, based on the idea that exerting control takes time during which no
other tasks can be executed, formulates that control exertion carries opportunity costs and
that quitting or forgoing control-greedy task could be a manifestation of the exploration-
exploitation trade-off.
(1) Limited intrinsic physiological resource
The idea of control relying on a depletable limited resource was first formulated in the
specific case of tasks requiring sustained attention (Posner and Snyder, 2004, reedition of
1975 book; Shiffrin and Schneider, 1977). This concept was later extended to cognitive control
in general, after several teams found that subjects perform less well at tasks requiring
cognitive control if they previously performed another control-greedy task (e.g. Baumeister et
al., 1998; Garrison et al., 2019; Hagger et al., 2010; Muraven and Baumeister, 2000; Radel et
al., 2019; Schmeichel, 2007). For example, Garrison and colleagues (2019) first asked people
to write an essay about their last holidays. One group had no extra assignment, while the
second could not use the letters A or E. Following that writing assignment, subjects performed
a Stroop task (see Figure 5C for example of a Stroop task design). Garrison and colleagues
32
report that the participants that had previously performed the controlled-writing task had
overall lower performance at the Stroop task. This phenomenon of lower control capacity
after control exertion is called ‘ego-depletion’ (phenomenon reviewed by Hagger et al., 2010),
and has been described by some as a ‘refractory period’ occurring after cognitive control
exertion (Baumeister et al., 1998). This phrasing suggests a structural, physiological constraint
on cognitive control capacity.
Shenhav et al., 2017, Figure 1
Figure 13 : Schematic representation of the different constraints on control exertion that could
account for limited control capacities and control aversiveness (adapted from Shenhav et al.,
2017).
The example situation proposed is the case of a driver, attending multiple information (e.g.
visual, auditory) in an effort to drive safely and reach his destination. Metabolic accounts of
the limitations for control investment state that control consumes and therefore depletes a
metabolic resource critical for neuronal activity, glucose or glycogen. This account is shown in
green. Control-capacity hypotheses of control limitation state that the control that can be
applied is limited (due for example to working memory or information processing capacity
limitations) (blue) and interference-based accounts of cognitive control limitations state that
the system limits the number of ongoing processes to reduce the risk of interferences (orange).
As suggested by the figure, these theories are not mutually exclusive, but may constitute pieces
of a larger scheme.
The hypothesis that the limited resource for mental effort is metabolic was formulated
by Baumeister and colleagues in the late 1990s (Baumeister et al., 1998; Baumeister and
Heatherton, 1996; Muraven et al., 1998). Cerebral activity consumes glucose (Benton et al.,
1994; Fairclough and Houston, 2004; McNay et al., 2001, 2000). Although the brain is
continuously supplied with glucose by the blood flow, glucose can be consumed faster than it
is supplied, thus depleting active brain areas in metabolic substrate (Benton et al., 1996;
Fairclough and Houston, 2004; McNay et al., 2001). The energy management system of the
brain is organized around the neuro-glio-vascular ensemble, capable of adapting rapidly to
fast increases in metabolic demand. This metabolic system is organized around three major
33
metabolites: glucose, glycogen and lactate (Dienel, 2017; Gruetter, 2003; Hertz et al., 2007;
Hui et al., 2017; Müller, 2014; Pellerin and Magistretti, 1994; Walls et al., 2009). Glycogen is
stored quasi exclusively in astrocytes (Magistretti et al., 1981; Phelps, 1972; Watanabe and
Passonneau, 1973), representing, in glucose equivalents, more than all the cytosolic glucose
of neurons (Gruetter, 2003). Glycogen can be synthesized in periods of low brain activity (e.g.
sleep, Swanson, 1992) or high glycaemia (Nelson et al., 1968) to store excess glucose, while in
periods of high activity, it is depolymerized locally to form lactate (another metabolic
substrate for neurons), thus providing a substrate for the increased metabolism (Gailliot,
2008; Swanson et al., 1992). In that sense, astrocytic glycogen acts as a buffer, an energy
reservoir. In conditions of hypoglycemia, astrocytic glycogen can support brain activity for
more than an hour (Gruetter, 2003). Note that besides serving as glucose storage for
emergency situations, glycogen is in fact part of a dynamic system, with a variable proportion
of glucose systematically stored in the form of glycogen before becoming available to the
neurons via glycogenolysis. This mechanism is referred to as ‘glycogen shunt’ (Obel et al.,
2012; Shulman et al., 2001; Walls et al., 2009).
Blood glucose was first proposed as the depletable resource that cognitive control
aversion aims at sparing. Indeed, low blood glucose has well-described detrimental effects on
many mental processes (e.g. language, visual processing), for instance cognitive control.
Conversely, glucose intake has been shown to reduce ego-depletion effects following control
exertion, suggesting that it allowed compensating the putative depletion of resources. A large
part of the research focusing on the relationship between blood glucose and cognitive
performance was conducted in diabetic individuals, in whom dramatically low glycaemia can
occur as a result of excessive insulin administration, but also in healthy subjects after a short
period of fasting (Kennedy and Scholey, 2000; Benton et al., 1987; Donohoe and Benton, 1999;
Fairclough and Houston, 2004; reviewed by Feldman and Barshi, 2007, and Gailliot and
Baumeister, 2007).
For example, Gailliot and colleagues (2007) showed that, following a control-greedy
task, subjects’ glycaemia decreased, and subjects with low glycaemia before the task had
lower overall performance at control-greedy tasks. Moreover, the ego-depletion
phenomenon observed after subjects had been asked to perform control-greedy tasks for a
prolonged time was annihilated by glucose-drink consumption (Figure 14). Note that these
authors gave artificially sweetened drinks (zeros calories) to the group who did not receive
34
sugar, and controlled for the effect of drink liking on performance depletion, in an effort to
eliminate the potential confound of the sweet drink serving as an incentive for control
allocation.
Gailiot et al., 2007, Figure 2
Figure 14 : Ego-depletion after a controlled attention task, in a Stroop task with or without
glucose consumption (adapted from Gailliot et al., 2007).
Subjects were first instructed to watch a video, either without any specific assignment (“Watch
normally”, black bars), or with the instruction to focus only on the female character in the video
and not on words popping at the bottom of the screen (selective attention task, “Control
attention”, grey bars). After completing that task, subjects either received a glucose-
sweetened lemonade or an aspartame-sweetened lemonade and rated the likeability of their
drink. They then performed a Stroop task. Performance in the group receiving no glucose
followed an ego-depletion pattern: performance was lower after the high-demand task. This
ego-depletion effect was not present in the group receiving glucose.
The effects of blood glucose and glucose administration on cognitive performance,
however, have been widely discussed, especially due to repeated failures to replicate findings
from Gaillot and colleagues (2007) regarding decreases of blood glucose following control-
greedy tasks (Dvorak and Simons, 2009; Molden et al., 2012; reviewed by Kurzban, 2010).
Moreover, even if the findings from Gaillot and colleagues were to be replicated, the size of
the decrease of the glycaemia appears to be anecdotal and detractors of the theory argue that
it would be too small to affect cognitive processes (Christie and Schrater, 2015). Findings
regarding the enhancement of cognitive control following glucose administration have, on the
contrary, been replicated in humans (Denson et al., 2010; Owen et al., 2012) and animals
(Miller et al., 2010) but are merely in favor of a beneficial effect of glucose on control
allocation (possibly as an incentive) without validating glucose as the limited resource for
control.
An alternative theory regarding the resources that organisms intend to spare by
avoiding cognitive control exertion concerns astrocytic glycogen (Christie and Schrater, 2015).
35
Christie and Schrater theorized that the brain plans and manages its metabolic resource
consumption and trades-off cognitive effort, and therefore increased performance, for
reductions in metabolic substrate consumption by reductions of neuronal activity. These
authors, however, consider astrocytic glycogen as the metabolic substrate that the system
consumes and aims at maintaining at a sustainable level (Figure 15). In brief, Christie and
Schrater devised a model that accounts for both cognitive control avoidance and ego-
depletion phenomena, in an effort to spare glycogen for later demanding tasks. Noticeably,
LC activation promotes short-term glycogenolysis (depolymerization of glycogen to form
lactate) and longer term glycogen synthesis in target areas, and such effects and their
implications for the relationship between LC activity and effort processing will be discussed
later in the manuscript (see part I.C.1 and III). The metabolic relevance, i.e. the feasibility that
glycogenolysis can respond rapidly and sufficiently to adjust to the increased metabolic
demand in case of augmented brain activity, signaled by certain neurotransmitters (including
noradrenaline from the LC, see part I.C.1) was later tested by Coggan and colleagues (2018).
Their study showed, using a custom-made computational model, that Christie and Schrater’s
theory would be compatible with what is currently known of the enzymatic cascades involved
in astrocytic and neuronal metabolism.
(A) Christie and Schrater, 2015, Figure 1A (B) Christie and Schrater, 2015, Figure 1B
Figure 15 : Glycogen as a buffer for brain metabolic resources (adapted from Christie and
Schrater, 2015)
(A) Schematic representation of the reactions transforming blood glucose from capillaries into
glycogen, stored in astrocytes, later depolymerized to form lactate (another metabolic
substrate for neurons), and released to neurons in case of high metabolic demand (G-6-
P=glucose-6-phosphate). The mechanism of systematic capture and polymerization of glucose
into glycogen by astrocytes is referred to as the glycogen shunt.
(B) Energy flow in the model from Christie and Schrater (GLUT1 and 3: glucose transporters)
Note also that Holroyd (2016) has evoked the possibility that the aversiveness of
cognitive control could be associated to accumulation of waste (specifically amyloid-β protein
36
in the interstitial fluid) due to a saturation of the waste clearing system, rather than
consumption of resources. This possibility is complementary to the resource account of
mental effort aversiveness and will not be detailed here.
To summarize, several teams have proposed metabolic accounts for the aversiveness
of cognitive control exertion. Specifically, avoiding control would allow sparing glucose or
glycogen, or could be avoiding accumulation of metabolic waste. The hypothesis that
astrocytic glycogen could be the limited resource for cognitive control investment is of
particular interest for this manuscript considering the relationship between LC activity and
glycogen metabolism (see part I.C.1).
(2) Limited computational resource
Some authors have suggested that the limited resource for cognitive control could be
computational. Working memory capacity has been proposed as limiting (Anderson, 1983;
Cohen et al., 1990), as tasks requiring cognitive control very often involve working memory (if
only to recall task rules and information), and working memory capacity is known to be limited
(Cowan, 2005; Luck and Vogel, 1997; Miller, 1956). Only indirect proof can however be
provided to support that hypothesis. Moreover, limitations in working memory capacity can
hardly account for long-term depletion in control capacity after control exertion, and this
frame mostly accounts for the difficulty to conduct several tasks requiring working memory
simultaneously, leading this interpretation to be criticized (Shenhav et al., 2017).
Generalizing this theoretical frame, the multiple resource hypothesis states that
limitations in control exertion are the consequence of bottlenecks in processing (Allport, 1954;
Logan, 1985). For instance, one is not capable of resolving two mathematical equations at the
very same time, either due to intrinsic incapacity of the system to conduct two operations at
the same time, or due to an effort of the system to avoid any risk of crosstalk between
concurrent tasks. Just like working memory limitation, bottlenecks in processing mainly justify
that two control-demanding tasks usually cannot be conducted at the same time.
Nonetheless, this incapacity to efficiently conduct several demanding tasks at the same time
due to computational limitations would lead allocating control to one specific task at a given
time to carry opportunity costs (see below, part I.A.2.e)(3)), which could account for the
aversiveness of control.
37
To summarize, limitations in control exertion have been hypothesized to be caused by
computational limitations. Although this point of view can account for the difficulty or even
incapacity of multitasking, it can hardly account for ego-depletion phenomena. This does not
make this theory invalid, but merely insufficient, and the existence of bottlenecks in
processing could be part of the reasons for limited control exertion.
(3) Motivational resource: exploration-exploitation trade-
off and opportunity costs of control
In contradiction, or at least as a complement to the theories presenting control as a
process consuming limited depletable physiological or computational resources, several
authors have conversely proposed that depletion effects could be accounted for by
motivational mechanisms. Still, the motivational and metabolic accounts of cognitive control
aversion and depletion may not be incompatible but merely “components of a more general
framework” (Christie and Schrater, 2015; Kurzban et al., 2013).
The motivational hypothesis emerges from the findings that offering rewards (Boksem
et al., 2006; Inzlicht et al., 2014; Muraven and Slessareva, 2003) as well as other manipulations
(e.g. self-conviction of unlimited willpower, Job et al., 2010) can reduce and even eliminate
depletion effects. Moreover, theories of control allocation (expected value of control theory,
see part I.A.2.d)(3) state that subjects make a trade-off between cognitive control and
accuracy, defined as the maximization of the expected value (expected reward minus
expected effort). In that frame, decisions to explore, e.g. pick a new unknown option, would
rarely be categorized as accurate. Indeed, since the expected utility (expected reward and
required effort) of the new option (exploration) is usually highly uncertain, it is down-weighed
by its risk (Dow and Werlang, 1992), compared to the well-known option (exploitation). Since
permanent exploitation is maladaptive (Stephens and Krebs, 1986), some teams have
suggested that the necessity to explore could be why control exertion is costly and limited.
Specifically, Inzlicht and colleagues (Inzlicht et al., 2014; Inzlicht and Schmeichel, 2012)
theorized that the ‘refractory period’ following control-greedy tasks could be related to the
‘evolutionary pressure motivating organisms to balance their desires for exploitation versus
exploration’, causing a switch of motivational priorities to make the most of a given
38
environment. Pursuing control-greedy strategies, just like any exploration strategy, is
associated to opportunity costs. While focusing on a cognitive task, subjects cannot pursue
other courses of action that might pertain better rewards, i.e. they cannot explore. Such
opportunity costs are especially true for decision-making considering that decision strategies
involving high control allocation to ensure high accuracy tend to be associated with increased
time spent deciding, less decisions per unit of time, and therefore less potential rewards per
unit of time. Humans and animals have indeed long been known to behave according to a
speed-accuracy trade-off, i.e. they routinely do not spend as much time making a decision as
could be required to be perfectly accurate (Chittka et al., 2003; Franks et al., 2003; Heitz and
Schall, 2012; Wickelgren, 1977; reviewed by Heitz, 2014).
To summarize, motivational accounts of the limitations to control exertion state that
allocating high cognitive control is associated to opportunity costs, and these opportunity
costs are believed to constitute a decision variable for the determination of the amount of
control worth investing in a task (Kurzban et al., 2013). Dynamically balancing exploitation and
exploration, i.e. behavioral flexibility, implies regulating the extent to which one engages in
tasks, i.e. exerts cognitive control, as opposed to disengaging from the task to sample other
opportunities. Note that several recent studies on behavioral flexibility have pointed towards
an implication of the LC and of noradrenaline (NA). This literature will be reviewed further
(see part I.B.4).
f) Behavioral proxies for mental effort
In order to be able to access the amount of cognitive control exerted by subjects in
laboratory tasks, identifying reliable proxies for it is crucial. Indeed, a possibility would be
repeatedly asking subjects to rate their mental effort levels, but it would be time consuming,
and it is obviously not feasible in the case of animal research. In that part, I therefore examine
two behavioral parameters, RT and performance, that can be used to approximate how much
effort was invested by subjects.
(1) Reaction time
Everyday life cognitive operations, calculations for example, provide trivial examples
of the correlation between cognitive task difficulty, and duration of the task. For instance,
computing a one-digit multiplication in one’s head will take considerably less time than
39
computing a four-digit multiplication. Formally, this suggests that RT, i.e. the time elapsing
from the onset of a cognitively demanding operation until its end, increases with cognitive
demand. Under the assumption that control exertion is relatively constant over time, RT,
representing the duration of the cognitive effort, can serve as a satisfying proxy for the total
cognitive control allocated. This constitutes a simplification, as Shenhav and colleagues (2013)
did theorize that the intensity of cognitive control can vary across situations (tasks and trials).
In the 1970s, Wright, Bettman and Zins (Bettman and Zins, 1979; Wright, 1975)
provided some indirect account that more effortful strategies to solve problems tend to be
associated with longer RTs. They asked subjects to solve problems using different sets of rules,
differing in the level of mental effort required. The rule that subjects reported to be the least
effortful to apply corresponded to shorter RTs. However, no mechanistic account of this
relationship was formulated at the time.
Bettman, Payne and Johnson (1990) later proposed a theoretical framework to explain
such findings, summarizing and generalizing the theoretical work of several other teams
(Huber, 1980; Johnson and Payne, 1985; Newell and Simon, 1972; Payne et al., 1988; Shugan,
1980). They described decision-making as a succession of elementary information processing
events, and different decision-making strategies for the same decision could include more or
less of these elementary processes. For example, when asking subjects to choose between
two alternative courses of action, each action associated to a certain required effort and a
certain reward, a potential strategy would be choosing based on the rewards only, as opposed
to a strategy involving considering and weighing both dimensions. The former strategy will
trivially require less computational steps compared to the latter: to compare only rewards,
the subjects only have to compute a subtraction (one step), while to compare options based
in both attributes, they would for example have to compute the value of each option then
subtract these values (three steps). Intuitively, the former strategy, since it includes less steps,
will correspond to less cognitive effort compared to the latter. Experimentally, several studies
demonstrated a positive link between the putative number of elementary information
processing steps in a task and RT (Bettman et al., 1990; Carpenter and Just, 1975), as well as
a negative effect of the number of such steps on subjects’ choices of strategies to solve a task
(Bettman et al., 1990; Payne et al., 1988).
In the specific case of value-based decisions, RT is hypothesized to be decomposed into
at least three independent components: time necessary for stimulus processing, for sufficient
40
evidence accumulation to reach a decision, for triggering chosen action (Dmochowski and
Norcia, 2015). Over the past two decades, the drift-diffusion model (DDM), a sequential
sampling model, has become the preponderant model used to fit behavior in value-based (as
well as perceptual) decision-making (e.g. Bogacz et al., 2006; Ratcliff and McKoon, 2008). This
model assumes that decisions consist in the accumulation of evidence favoring one alternative
over the other. In the DDM, the three stages of the RT are sequential and therefore their
added durations form RT.
The DDM predicts that the greater the difference in value for the two options, the
shorter the RT, since the accumulation of evidence reaches a decision bound faster. Behavioral
studies have confirmed this finding (Hunt et al., 2012; Krajbich et al., 2010; Krajbich and
Rangel, 2011; Milosavljevic et al., 2010; Morris et al., 2019; Philiastides and Ratcliff, 2013).
Note that accuracy (proportion of trials in which subjects chose the best option, based on their
previous subjective ratings) increased with value difference, thus indicating that subjects
found choices with high value difference easier (Milosavljevic et al., 2010). This also confirms
that extensive information processing is perceived as mental effort (Anderson, 1990; Bettman
et al., 1990; Simon, 1956; Smith and Walker, 1993). Subjects therefore probably invested more
cognitive control to make the choice in case of low value difference, and this excess in control
exertion resulted in increased time spent to decide (increased RT).
Besides this effect of value difference on RT, a second interesting effect was observed
when examining RT variations when the overall value of the trial increased, i.e. when the sum
of the subjective values of the options presented increased. Hunt and colleagues (2012) report
that subjects were faster (negative effect on RT) when the overall value of the trial increased,
over and above the effect of value difference (Figure 16B). In coherence with that finding,
McGuire and Botvinick (2010) showed a positive relationship between mean RT throughout
blocks of trials and avoidance ratings of the whole block (Figure 16E). Cognitive control is
required to override default courses of action (Epstein, 1994; Sloman, 1996). In the case of
trials with low overall value or high avoidance ratings, the default response would have been
refusing the trial, which is not an option in that task. Consequently, subjects probably
mobilized more cognitive control to perform the trial in low overall value conditions.
Indirectly, the negative relationship between overall value and RT thus suggests that RT could
serve as a proxy for cognitive control in that case.
41
One limitation of the use of RT as a proxy for mental effort is the risk of circular
reasoning (Navon, 1984; Shenhav et al., 2017). One example for this circularity can be found
in the Stroop task, described above (Figure 5C). RTs are known to be in average longer in the
incongruent conditions, compared to the congruent conditions, and this difference is referred
to as the ‘Stroop effect’. In the theoretical frame that RTs correlate positively with the mental
effort invested in tasks, one would conclude that RTs generally increase in the incongruent
conditions because the subjects invest more effort in identifying the correct answer, as
distractor presence increases the difficulty. Another possibility could however be that, when
difficulty increases, subjects do not increase the invested effort sufficiently, and therefore take
longer to respond. In that case, RT would correlate with the subjective difficulty of the trial
rather than the effort invested.
(A) Hunt et al., 2012, Figure 2 (B) Hunt et al., 2012, Figure 3A (C) Hunt et al., 2012, Figure 3B
(D) Kool and Botvinick, 2018, Figure 1A (E) McGuire and Botvinick, 2010, Figure S1
Figure 16: Influence of value difference (VD) and trial overall value (OV) or avoidance ratings
and on RTs in a value-based decision-making task
(A) Task design. Subjects were instructed to choose between two options represented by
rectangles, with different reward sizes (rectangle size) and probabilities (percentage written
below). After choosing, they received a feedback on the rewards of the options (red for non-
rewarded, green for rewarded) (adapted from Hunt et al., 2012).
(B) RTs as a function of value difference (VD, black) and overall value (OV, red). RTs decreased
with both increasing VD and OV (adapted from Hunt et al., 2012).
(C) Size of the negative linear regression coefficients for value difference (VD) and overall value
(OV) on RTs. Both had significant negative effects on RT (adapted from Hunt et al., 2012).
(D) Task design from McGuire and Botvinick, 2010. Subjects were repetitively asked to choose
between pairs of previously unknown cues. One cue was associated to a low demand task, the
other to a high demand task (see Figure 7 for extra details, adapted from Kool and Botvinick,
2018).
(E) RTs as a function of trial number within block. Participants were told that trial blocks were
drawn from one of four ‘sources’. After each run, subjects rated their willingness to avoid
future blocks from the same ‘source’, thus providing post-hoc subjective mental effort ratings
42
for each cue. RTs are plotted by avoidance rating: level 1 in green, 2 in grey and 3 in red. Mean
RT is higher in high avoidance blocks (adapted from McGuire and Botvinick, 2010).
(2) Performance
Another potential proxy for the cognitive control involved in a task could be
performance, although it carries strong limitations. A first caveat lies in the definition of
performance. For instance, in a Stroop task, performance can be defined as the success rate
in giving the correct answer, but also as the inverse of the difference in RT between congruent
and incongruent conditions, as exerting cognitive control should reduce this difference.
Additionally, some authors argue that, under the condition that the subjects’ intrinsic
capacities are not exceeded and that they exert adequate cognitive control relative to the task
difficulty, increased effort should always allow maintaining performance when task difficulty
increases (Norman and Bobrow, 1975). In that frame, performance would not vary, thus
making it an inadequate proxy for the effort exerted (Backs and Seljos, 1994). Such a situation
is however merely theoretical, as studies using standard mental effort experimental
paradigms usually report differences in performance across difficulty levels, even when
participants do report exerting increased mental effort in the higher difficulty levels. This is in
line with Shenhav’s EVC theory (2017, 2013) that states that subjects do not systematically
engage the level of effort that would be required to absolutely maximize reward rate, but
compute a tradeoff between the cost of the effort engaged and the reward obtainable
through that effort, such that the EVC is maximized (see part I.A.2.d)(3)). In this frame,
performance is not a satisfying proxy for the cognitive effort. Again, using performance as a
proxy for mental effort is at risk of circular reasoning.
To summarize, behavioral proxies for cognitive control exertion have been proposed.
Performance appears quite imperfect, as it results from the interaction between subjective
difficulty and control exertion. Additionally, it can allow evaluating difficulty for groups of
trials, but not trial-by-trial. RT, conversely, can be measured trial-by-trial. It probably
correlates with the subjective difficulty of trials, but interpretations in terms of exerted
cognitive control must be taken with caution as they risk circularity. A way to solve the
circularity encountered with using RT and performance as proxies would be to introduce
mental effort exertion self-reports throughout the task. Another possibility would be assessing
43
mental effort through an independent measure, for instance a physiological measure. I review
the various measures proposed in the literature in the following part.
g) Physiological proxies for mental effort:
cardiovascular, metabolic and vegetative measures
Several teams over the past 60 years have investigated peripheral indicators of mental
effort: metabolic, cardiovascular and vegetative markers. I will review briefly these proxies.
(1) Cardio-vascular and metabolic proxies
Following early reports of the relation between heart rate and psychological tasks
(Lacey, 1967), several studies on mental load have reported increases in tonic heart rate.
These studies originally reported increases beyond what would have been expected from the
current physical (somatic) activity, and specifically beyond the concurrently measured
peripheral oxygen consumption (outside of the tasks, Carroll et al., 1986a; Turner and Carroll,
1985). These increases in heart rate were initially referred to as ‘metabolically exaggerated
heart rate’ or ‘additional heart rate’. In the same period, increases in energy expenditure
(oxygen consumption and carbon dioxide production) in response to mental effort (Carroll et
al., 1986a; Sims et al., 1988; Sims and Carroll, 1990; Turner et al., 1983; Turner and Carroll,
1985) were reported.
These findings were initially interpreted in terms of energy expenditure due to the
somatic demands associated to the mental effort task (e.g. moving a video-game joystick, Sims
et al., 1988; Turner et al., 1983; Turner and Carroll, 1985). Backs and Ryan (1992) however
ruled out such interpretation by having subjects perform a continuous working memory task
with two difficulty levels but equal motor demands. Participants were presented with a
succession of visual targets, and had to count the number of presentations of either one or
three targets. The study reports a small consistent increase in heart rate and oxygen
consumption, and faster and shallower breathing with increasing task difficulty. Backs and
Seljos (1994) completed this result by demonstrating that these metabolic and heart rate
increases were not dependent on differences in baseline metabolism and heart rate
(specifically across male and female subjects). Conversely, individuals with lower working
memory performance, and therefore perceiving the task as subjectively more difficult (as
44
shown by Ericsson and Simon, 1980), displayed stronger cardiovascular and metabolic
adaptations with task difficulty.
Furthermore, several have tried to refine the relationship between heart rate and
mental effort. Specifically, they decomposed the variations of heart rate by frequency (power
spectral analysis), and found a relationship between the 0.1Hz component of its variability and
several correlates of mental effort: working memory load (Aasman et al., 1987; Mulder and
Mulder, 1981), complex decision-making (Tattersall and Hockey, 1995), and variations in task
difficulty (Nickel and Nachreiner, 2003; Wilson, 1992; Jorna, 1992). Note that this frequency
component is supposed to be related to parasympathetic activity (Mulder, 1986, 1979).
Noticeably, differences in cardiovascular and metabolic activity across difficulty levels
were only reported in tasks requiring sustained mental effort (e.g. continuous working
memory task). They were not found in tasks requiring discrete computations separated by
relatively long time inter-trial-intervals (e.g. arithmetical task, Carroll et al., 1986a, 1986b),
although these tasks elicited overall increased heart rate and oxygen consumption. This
difference may merely reflect slow metabolic and cardiovascular adaptations or may be due
to an interaction between difficulty and time on task.
Overall, these early reports indicate that mental effort is associated to increased
metabolic and cardiovascular demand. These measurements are however imperfect, as
mental load, performance, physiological correlates of mental effort and effort ratings are only
correlated, and each proxy for mental load does not vary with the same components of mental
effort and with identical sensitivity (Sanders, 1979). Moreover, those findings do not allow
confirming the causality link between such cardiovascular and metabolic adaptations and
mental effort, since heart rate could increase in order to provide sufficient oxygen to the brain
regions involved in the current effortful task (direct pathway), or because effort requires
increased sympathetic tonus, which coincidentally increases heart rate and metabolic activity.
I will review the literature on the relationship between one component of the sympathetic
system, brain NA, and mental effort further (see part I.B.4).
(2) Pupil dilation
Another frequently used proxy for mental load is the task-evoked pupillary response,
i.e. the pupil diameter change in response to task stimuli or during task-related computations,
45
relative to a pre-stimulus baseline period (reviewed in Beatty, 1982; van der Wel and van
Steenbergen, 2018). This relationship is of particular interest for this manuscript considering
the well described positive correlation between LC activity and pupil diameter (Joshi et al.,
2016; Murphy et al., 2014; Varazzani et al., 2015), which will be reviewed later (see part
I.B.2.b).
The positive correlation between pupil diameter and mental activity is easily
observable in everyday life, as explained by Kahneman (1973):
“Face a mirror, look at your eyes and invent a mathematical problem, such as 81 times
17. Try to solve the problem and watch your pupil at the same time, a rather difficult exercise
in divided attention. After a few attempts, almost everyone is able to observe the pupillary
dilation that accompanies mental effort.”
First descriptions more than fifty years ago of pupil diameter correlating with mental
effort focused on tasks involving memory load (Kahneman and Beatty, 1966), arithmetical
problem solving (Hess and Polt, 1964) and decision-making (Simpson and Hale, 1969). In
recent years, pupillometry has become widely used in cognitive neuroscience studies, due to
its simplicity of use and to the low price of the equipment, and many teams have conducted
studies about the relation between event-related pupil dilation and mental effort. Most
studies, though, report a positive correlation between task difficulty and pupil dilation, but
cannot effectively differentiate between the difficulty of the task and the effort invested. For
example, several have used keep-track tasks, consisting in asking subjects to track a subset of
pre-assigned dots among a group of randomly moving dots (Figure 17A). This task is notably
more difficult when the number of dots that have to be tracked increases, and pupil dilation
during tracking has been reported to increase for increase numbers of dots to track, as shown
in Figure 17B (Alnæs et al., 2014; Wahn et al., 2016; Wright et al., 2013).
(A) Alnaes et al., 2014; Figure 1 (B) Alnaes et al., 2014, Figure 3
Figure 17 : The multiple-object tracking task (adapted from Alnæs et al., 2014).
(A) Task design: At the beginning of trials, subjects were presented with a certain number of
blue dots. Then a subset of the dots turned red (three here) for a short amount of time before
46
turning blue again. These dots would have to be tracked by the subjects throughout the trial.
During the rest of the trial, all the dots moved randomly (tracking period). At the end of the
tracking period, subjects had to report the position of the tracked dots by clicking on them.
Task difficulty increased with the number of dots tracked.
(B) Pupil time series in the multiple-object tracking task. Trials are separated by difficulty, i.e.
by number of objects to track. The x-axis represents the time from tracking onset (dots start
moving) and the y-axis represents pupil size, in mm. Pupil diameter was larger during tracking
period when the number of tracked objects was higher, i.e. task difficulty increased.
Note that several studies have additionally reported a positive relationship between
global task performance and baseline (outside of task) pupil diameter (Heitz et al., 2007; Meer
et al., 2010; Tsukahara et al., 2016). These authors argue that global task performance can be
a proxy for the intelligence of individuals in the domain tested (e.g. fluid intelligence, the
capacity to reason, independently of learned skills).
(3) EEG proxies for cognitive control
Electroencephalograms (EEG) are a non-invasive method of exploring brain activity
through electrodes placed in strategic positions on the skin of the head. Event-related
potentials (ERPs), stereotyped responses to stimuli recorded in EEG, can be decomposed in
several sub-components (Nidal and Malik, 2014), and the amplitude of two of them has been
shown to correlate with the cognitive control exerted: N2 and P3 (Figure 18). The N2
component has been shown to be related to conflict monitoring. Indeed, in Go/No-Go tasks,
the N2 is enhanced during No-Go trials compared to Go trials (Clayson and Larson, 2013;
Falkenstein et al., 1999; Folstein and Van Petten, 2008; Garavan et al., 1999; Tillman and
Wiens, 2011) and larger in subjects making fewer false alarm errors (Falkenstein et al., 1999).
P3 amplitude has also been linked to inhibitory control (Garavan et al., 1999). Additionally,
several studies established a link between N2 and P3 and working memory load (e.g. Pratt et
al., 2011; Qi et al., 2014; Wei and Zhou, 2020), and with task-switching costs (e.g. Gajewski
and Falkenstein, 2011; Jamadar et al., 2015; Kopp et al., 2020). Overall, these results indicate
that certain components of ERPs could serve as proxies for the perceived cost of the mental
effort or the allocated cognitive control.
47
To summarize, cognitive control exertion has been linked to increased heart rate,
oxygen consumption, pupil dilation (baseline and in response to task events) and amplitude
in two components of EEG ERPs, N2 and P3. Pupil diameter and EEG ERPs are of particular
interest because, first, they can be measured on a trial-by trial basis, and because pupil
diameter and P3 amplitude have been linked to LC activity (see, respectively, parts I.B.2.b) and
I.B.2.a) for details).
Figure from: Syntax in a blink: Early and automatic processing of syntactic rules as
revealed by event-related brain potentials. Dissertation, 172 S., Max Planck Institute for
Human Cognitive and Brain Sciences, Leipzig (2008).
Figure 18: ERP recorded in EEG. ERPs can be decomposed into several subcomponents: N1
(positive), P2 (negative), N2 (positive) and P3 (negative) 3
In brief, mental effort exertion, in the recent years, has been unified under the concept
of cognitive control. Cognitive control is required in tasks involving updating information in
working memory, switching between tasks and rules, and inhibiting automatic or default
responses. The brain system organizing cognitive control allocation has been theorized to
include three sub-units: a regulation unit implementing the control signal, defined by its
identity and intensity, a specification unit deciding on the control signal, and a monitoring unit
receiving feedback on the process of control exertion and consequently adapting the decision
parameters of the specification unit. Cognitive control is involved in decision-making,
especially (if not exclusively) if extensive processing of options is required. It also serves as a
decision variable, since exerting control is perceived as aversive and tends to be avoided.
Models of decisions regarding control exertion have been formalized. The aversiveness of
mental effort is believed to reflect a mechanism of resource sparing, although the exact nature
of such resources has been debated. They could be metabolic (blood glucose or astrocytic
3 Figure from: Syntax in a blink: Early and automatic processing of syntactic rules as revealed by event-related
brain potentials. Dissertation, 172 S., Max Planck Institute for Human Cognitive and Brain Sciences, Leipzig
(2008).
48
glycogen), computational (bottlenecks in processing) or motivational (opportunity costs and
exploration-exploitation tradeoff). Measuring mental effort is difficult, but some indirect
proxies have been proposed: behavioral proxies, RT and performance, and physiological
proxies, heart rate, oxygen consumption, pupil dilation, and magnitude of EEG ERPs N2 and
P3.
3. Mental and physical effort, evidence for common resources
and partially common processing systems
“Mental effort is like physical effort—people dislike both.” Camerer and Hogarth, 1999
In the previous parts, I reviewed separately the characteristics of physical and mental
effort. Strikingly, both categories of effort seem to share common characteristics. Although
subjective physical effort perception could, at first glance, be considered to relate solely to
muscular activity, it was shown to be independent from sensory feedback and dependent on
a duplication of the motor command in the SMA (see part I.A.1.b). Additionally, physical effort
perception increases after curare administration (Gallagher et al., 2001), i.e. when extra
neuronal stimulation from the motor cortex to the muscles is required, suggesting that it is
the intensification of the neuronal activity required for the muscular contraction that is
perceived as effortful, not the muscular contraction itself.
Furthermore, mental fatigue has been repeatedly shown to impact physical effort
performance, thus generalizing the concept of ego-depletion (e.g. Boksem and Tops, 2008;
Martin et al., 2018; McMorris, 2020; Schiphof-Godart et al., 2018; Van Cutsem et al., 2017;
reviewed by Habay et al., 2021). For example, you may find it difficult to motivate yourself to
go for a run tonight after making the mental effort of reading this manuscript. At the
motivational level, a shift in the perception of physical effort following mental effort has been
suggested (McMorris, 2020). Along the same lines, excessive training, both in intensity and
frequency, can lead athletes to experience overtraining syndromes (Cadegiani and Kater,
2019; Meeusen et al., 2013; Smit, 1978). The major symptoms of overtraining syndrome
include (but are not restricted to) depression and low motivation (Meeusen et al., 2013),
indirectly suggesting that excessive physical effort exertion could lead to depleting a putative
critical resource for cognitive control.
49
At the cognitive level, one potential overarching cognitive explanation for such
relationship between cognitive control and physical effort exertion stems from the strong
possibility that physical effort requires exerting cognitive control, at least to decide to engage
in physical effort (or to decide to maintain it in the case of endurance training). Cognitive
control is required to override default responses (Epstein, 1994; Sloman, 1996). When facing
a decision whether to engage in a course of action requiring physical effort (or whether to
continue the ongoing physical effort in the case of endurance tasks), the default response
would be forgoing the effort, considering the aversive nature of physical effort (see part
I.A.1.a). For example, going for a run requires cognitive control to overcome the desire to stay
at home and rest. In that frame, mental and physical effort tasks may share (at least partly)
common resources, processing mechanism and neuronal substrates.
The physiological mechanism behind this impact of mental fatigue on physical effort
capacity and vice versa is, however, still debated. Some have proposed an accumulation of
adenosine (metabolic waste) in the brain, and in particular the anterior cingulate cortex (ACC)
(Pageaux et al., 2014). This idea is debated since increases in brain adenosine after physical
effort are short-lasted, while cognitive effects would require long-lasting increases in
adenosine (McMorris, 2020). Another possible mechanistic account, although not formally
demonstrated, comes from the combination of the hypothesis from Christie and Schrater
(2015, see part I.A.2.e)(1) that the depletable resource for cognitive control could be astrocytic
glycogen, combined with the result that physical effort depletes astrocytic glycogen (Matsui
et al., 2017). Together, these points would suggest that cognitive and physical effort could be
drawing on similar resources in the brain. This manuscript specifically aims at exploring the
specific role of the LC/NA system in effort processing, hypothesizing that it participates in the
processing of both (see part I.B.4).
50
B. Role of the locus coeruleus in effort processing:
review of the literature
The first part of this manuscript summarized the literature regarding physical and
mental effort at the conceptual level, their impact on decisions, and proposed proxies for
evaluating them. Additionally, clues for a potential common processing system for both
mental and physical effort were proposed. The central part of the work of this thesis concerns
the neurophysiology of the LC, and its involvement in effort processing in general. This part
will consist, first, of a brief review of the literature concerning the anatomy, histology and
physiology of the LC, and frequently used proxies for LC activity, followed by a part focusing
on accounts converging towards an implication of the LC in effort processing. Specifically, I will
review evidence regarding the implication in three processes involving resource mobilization,
arousal, stress and attention. Then, I will review general evidence for a role of the LC/NA
system in the execution of costly goal-directed actions, before focusing on the particular case
of cognitive control exertion for decision-making and its consequences. In a last subpart, I will
examine potential mechanisms for the effects of LC activity on cognitive processes, and
especially on cognitive control exertion.
1. Anatomy, histology and physiology of the locus coeruleus
a) Anatomy and histology
The existence of the LC has been known since the XVIIIth century, although it was
Dahlstroem and Fuxe (1964) who first identified that LC neurons produce monoamines, and
specifically NA. The LC is a small brainstem nucleus, constituted of quite large neurons
containing neuromelanin (Nieuwenhuys et al., 2008; Olszewski and Baxter, 1954), about
45000 to 50000 in humans, 10000 in rhesus macaques (Macaca mulatta, Sharma et al., 2010)
and 1500 to 2000 in rats and mice (Goldman and Coleman, 1981). It is located lateral to the
fourth ventricle, just under the anterior part of the cerebellum and the inferior colliculi (Aston-
Jones and Bloom, 1981; Bouret and Richmond, 2009; Grant et al., 1988; Kalwani et al., 2014,
Figure 19). Note that these landmarks are routinely used in magnetic resonance imaging (MRI)
51
and during electrophysiological recordings to locate the LC, since it cannot be visualized during
a routine MRI (MRI methods to image the LC are reviewed in Liu et al., 2017).
(A) Wikibrainstem, fibratlas.univ-tours.fr (B) Wikibrainstem, fibratlas.univ-tours.fr
(C) Grant et al., 1988, Figure 1
(D) Kalwani et al., 2014, Figure 5A (E) Kalwani et al., 2014, Figure 5B
(F) Kalwani et al., 2014, Figure 5F
Figure 19: Anatomy and histology of the LC in humans and monkeys
(A,B) Human MRI with T2 contrast indicating the position of the LC in red and the position of
nearby structures. (B) Nearby structure limits are highlighted with specific colors (anatomic T2
MRI, adapted)
(C) Histology of the LC in monkeys (Macaca arctoides). The position of the LC was marked after
recordings by iron deposits, made by passing anodal current when the electrode was located
immediately dorsal to the LC. These deposits (arrows), later marked with Prussian Blue coloring
are therefore just dorsal to the LC (adapted from Grant et al., 1988).
(D,E) Histology from the right hemisphere in a monkey (NB: the arrow shows a lesion from
electrode passage) (Adapted from Kalwani et al., 2014).
(F) Tyrosine hydroxylase immunohistochemistry allowing identifying NA producing neurons of
the LC in dark brown (Adapted from Kalwani et al., 2014).
Abbreviations:
4V and ven: Fourth ventricle; bc: brachium conjonctivium; 4x: trochlear decussation; Cb/Cb2,
cerebellum; CGPn: Central gray of the rhombencephalon; ctg: central tegmental tract; LC:
locus coeruleus; LV, lateral ventricle; mcp, middle cerebellar peduncle; Me5: mesencephalic
trigeminal nucleus; mlf, medial longitudinal fasciculus; MPB: medial parabrachial nucleus;
Mo5N: rostral tip of the motor trigeminal nucleus; Pn and PnO: pontine nuclei; scp : superior
cerebellar peduncle; subCD/subCV: dorsal/ventral parts of the subcoeruleus; vsc: ventral
spinocerebellar tract.
52
b) Electrophysiological properties
In vivo, as far as electrophysiology is concerned, LC neurons are characterized by low
rates of spontaneous activity, below 4Hz at rest and up to 5 to 6Hz during stress exposure,
broad waveforms (>0.6ms for the initial peak), and a stereotyped activation-pause responses
to brief unexpected stimuli (auditory or tactile stimuli, e.g. knock on door, see Figure 20). The
activation usually consists of a few spikes (sometimes only one), and neurons habituate fast
in case of repetition of the stimulus (Aston-Jones and Bloom, 1981; Bouret and Richmond,
2009; Bouret and Sara, 2004; Grant et al., 1988; Hervé-Minvielle and Sara, 1995).
Noradrenergic fibers emerging from the LC are not myelinated and therefore exhibit a slow
conduction velocity, ranging from 0.5 to 2.4m/s, with a mean of 0.9m/s (Aston-Jones et al.,
1985). As a consequence, the discharge of NA in target areas happens about 150ms from LC
neurons’ activation (30 to 135ms for the somatosensory cortex, 50 to 110ms for the frontal
cortex, 80 to 135ms for the occipital cortex, Aston-Jones et al., 1980; Berridge and
Waterhouse, 2003; Foote et al., 1983).
(A) Grant et al., 1988, Figure 7a (B) Grant et al., 1988, Figure 8a
(C) Kalwani et al., 2014, Figure 3H
Figure 20: LC neurons respond to surprising stimuli with a characteristic activation-inhibition
response
(A) Multi-unitary recording in the LC, peri-event time histogram. A surprising sound (black
arrow) produced a robust increase in rate (adapted from Grant et al., 1988).
(B) Single unit recording in the LC, around surprising sound (door knock). Top: event timing.
Bottom: spikes recorded from an LC unit, each vertical bar indicates a spike. The unit responded
with several spikes to the door knock (adapted from Grant et al., 1988).
(C) Multi-unitary recording in the LC, around unexpected auditory stimulus. Top: raster plot;
Bottom: peri-event time histogram. LC neurons responded with a burst consisting of a few
spikes around the onset of the beep, followed by an inhibition period (adapted from Kalwani
et al., 2014).
53
c) Connectivity to other brain structures
Recent studies on the anatomo-functional properties of the LC have benefited from
novel techniques allowing precise dissections of the internal anatomy and structure of the
connections of the LC with the rest of the brain. The most important recent finding regarding
the anatomy of the LC concerns its modular organization.
LC neurons receive diverse inputs from the rest of the brain (Aston-Jones et al., 1991;
Luppi et al., 1995; Schwarz et al., 2015). Viral tracing studies revealed that the LC receives
direct inputs from more than a hundred (111) regions of the brain, with individual LC cells
generally receiving inputs from nine to fifteen regions (Schwarz et al., 2015). In addition to
these various inputs, LC neurons may have diverse sensitivities to neurotransmitters as they
display a very variable pattern of expression of receptors. First, LC neurons are interconnected
and therefore sensitive to NA. However, adrenergic receptors are not evenly distributed
throughout the LC (Chamba et al., 1991). Overall, α2 receptors are predominant in quantity in
the whole LC, in addition to a minority of α1 receptors (Young and Kuhar, 1980), and these
two receptors are believed to have opposite effects on the firing rate of LC neurons (Szabadi,
2013). Importantly, α2 receptors were found to be more abundant in the posterior part of the
LC, while α1 receptors appeared to be mostly localized in the anterior part of the LC (Chamba
et al., 1991). Regarding sensitivity for other neurotransmitters, coming from axons projecting
from exterior brain areas, acetylcholine receptors, and in particular nicotinic receptors, are
not evenly distributed throughout the LC either, and such heterogeneity of distribution was
shown to be associated with heterogeneous excitatory response strength to acetylcholine
(Léna et al., 1999).
LC neurons are the sole source of NA, a catecholamine, for the whole forebrain and
the LC as a whole is virtually sending axons to all brain areas (Levitt and Moore, 1978; Morrison
et al., 1978). Note, though, that NA is not the only neurotransmitter produced in the LC, as
neuropeptide Y, galanin (Simpson et al., 1999; Takeuchi et al., 2016; Xu et al., 1998) and
several other neurotransmitters have been found. Co-release of these molecules with NA
could serve to modulate its effects in target areas (Tsuda et al., 1989). Not all LC neurons
produce all these neuropeptides, and small subpopulations of LC neurons producing certain
neurotransmitters can be identified (Sutin and Jacobowitz, 1991). Additionally, DA, the
precursor of NA, is synthesized by tyrosine hydroxylase not only in dopaminergic but also
noradrenergic neurons. Co-release with NA by LC terminals in the hippocampus and the
54
cerebral cortex and in particular the prefrontal cortex (PFC) has been demonstrated (Devoto
et al., 2005, 2001; Devoto and Flore, 2006; Kempadoo et al., 2016; McNamara and Dupret,
2017; Smith and Greene, 2012). LC release of dopamine (DA) accounts for around 73% of all
the dorsal hippocampal DA (Kempadoo et al., 2016).
Besides such neurotransmitter heterogeneity, LC projections onto the different brain
areas are not uniform. Agster and colleagues (2013) characterized the repartition of synaptic
boutons across the rat brain, finding greater density in frontal cortices than other cortices,
and in superficial than deep layers. Comparable results were obtained in primate brains (Lewis
and Morrison, 1989; Morrison and Foote, 1986). Complementarily, several studies have
brought to light the modular organization of the LC, using various techniques. First, messenger
ribonucleic acid (mRNA) sequencing consists in using viruses to inject specific mRNAs, called
barcode mRNAs, into precise subpopulations of neurons. The mRNAs are then transported to
the synaptic boutons of the axons. Identifying the final locations of these barcode mRNAs
allows identifying the projection patterns of the neurons in which the viruses were injected
(Kebschull et al., 2016). Second, viral-genetic tracing consists in injecting viruses into specific
brain regions. The viruses are modified to infect specific subpopulations of neurons (for
instance the LC) in genetically modified individuals (Cre mice) and cause the infected neurons
to express fluorescent proteins. In the case of anterograde viruses, the infection occurs at the
level of the synaptic bouton of neurons and the viral material transits towards the neuronal
body. The location of the infected cell bodies (e.g. of the LC) and therefore the inputs of the
area of interest can be revealed thanks to the fluorescent protein expression. The opposite is
true for retrograde viruses. This technique allows identifying the input and output circuitry of
areas of interest (Schwarz et al., 2015). Third, chemogenetics allow functional dissections of
neuronal subpopulations. For instance, to identify the role of projections of the LC onto certain
areas, retrograde viruses causing the expression of specific receptors (under a genetic
promoter present only in noradrenergic cells) are injected in the projection area of interest.
The retrograde virus thus causes the expression of the receptor specifically in the LC neurons
projecting onto the area of interest. The expressed receptor can bind with a certain drug, and
this binding causes either the activation or inhibition of the neurons expressing it. In behaving
animals, this method allows identifying the processes for which specific projections of the LC
are crucial (Hirschberg et al., 2017; Uematsu et al., 2017).
55
Overall, studies uding these novel methods showed that LC neurons are grouped in
modules, receiving specific inputs and providing target-specific outputs. First, neurons differ
in terms of connections (inputs and outputs), with, for example, neurons receiving inputs from
the amygdala virtually not projecting onto the medulla (Schwarz et al., 2015). Second, these
connectivity-defined modules tend to be morphologically different at the cell level, and
spatially grouped (especially along the antero-posterior axis, Loughlin et al., 1986). The dorsal
part of the LC is constituted mostly of fusiform cells while ventral LC neurons tend to be
multipolar (Loughlin et al., 1986; Swanson, 1976). Moreover, no medulla projecting neurons
can be found in the posterior part of the LC, and most forebrain-projecting LC neurons are
located dorsally. LC neurons projecting onto the cortex have, however, been found to be
scattered throughout the nucleus (Schwarz et al., 2015). Additionally, LC axons tend to
collateralize throughout the brain, i.e. individual LC neurons project onto several brain areas
(Nagai et al., 1981; Room et al., 1981; Steindler, 1981). Interestingly, Chandler and colleagues
(2014) demonstrated that such finding does not apply within cortical areas (orbitofrontal
cortex, anterior cingulate cortex, medial prefrontal cortex, and motor area 1). In other words,
LC neurons projecting onto a specific cortical area do not project onto another cortical area,
although they may still project both onto a cortical and a subcortical area (Uematsu et al.,
2017).
Third, different input and projection patterns are associated with differential
implication in cognitive processes, for example learning (Uematsu et al., 2017; Wagatsuma et
al., 2018), or nociception and aversion (Hirschberg et al., 2017). Despite focusing on restricted
subsets of cognitive functions and projection areas, these studies provided evidence of the
necessity to stop considering the LC as homogenous, anatomically and functionally. Early
studies reported that a small proportion of LC neurons exhibit different electrophysiological
properties ex vivo (brain slices, Williams et al., 1984) and in particular different shapes of
action potentials and different spontaneous firing rates. More recently, Totah and colleagues
(2018) recorded a very large number of LC single-units in vivo (anaesthetized rats) and
analyzed the relation between their spontaneous and evoked firing, their projection patterns,
and their the shape of their action potentials. They showed that LC neurons generally do not
fire synchronously (15 to 16% synchrony), although small ensembles of neurons, generally
spatially grouped, sharing gap junctions, and likely to project onto identical areas can be
synchronized. Additionally, the neurons of these ensembles tended to share waveform
56
characteristics, and waveforms differences were associated with differences across a wide
range of other electrophysiological properties. This study reinforced the idea that LC neurons
are not functionally homogenous, and that different electrophysiological properties tend to
be associated with different projection patterns, and therefore putatively different effects on
cognitive functions.
Overall, this body of literature provides evidence that LC neurons are not homogenous,
and are dividable into modules differing in their patterns of inputs and outputs, cell
morphology, electrophysiological properties, receptor expression pattern and
neurotransmitter expression.
d) Noradrenergic receptor distribution and actions in
target areas
Adding heterogeneity on top of the modular organization of the projections of LC
neurons, noradrenergic receptors are not evenly distributed throughout the brain. NA has
nine different receptors: α1A, α1B, α1D, α2A, α2B, α2C, β1, β2 and β3. α1 and β receptors are
located post-synaptically, while α2 receptors have both presynaptic and post synaptic
locations, with a majority located post-synaptically (U’Prichard et al., 1979). Adrenergic
receptors have different affinities for their ligand and distribution patterns throughout the
brain, with α2 adrenergic receptors displaying a much higher affinity for NA compared to the
others (Arnsten, 2000). Illustrating the uneven distribution of adrenergic receptors, higher
concentration of α2C receptors have been found in the deep cortical layers compared to the
superficial layers (Nicholas et al., 1993). Consequently, the effects of noradrenergic release at
the neuronal level is highly dependent on the brain area and on the current availability of NA
(reviewed by Arnsten, 2000). For instance, in the amygdala and hippocampus, fixation of NA
on β receptors (Cahill and McGaugh, 1996; Hopkins and Johnston, 1988; Roullet and Sara,
1998), in synergy with α1 receptors (Chaulk and Harley, 1998; Ferry et al., 1999; Puumala et
al., 1998) facilitates memory consolidation. Conversely, α2 receptor stimulation impairs it in
the amygdala (Genkova-Papazova et al., 1997) but not the hippocampus (Sirviö et al., 1991).
This pattern is reversed for the PFC. Indeed, β receptors do not have much effect on PFC-
associated functions such as working memory (Arnsten and Goldman-Rakic, 1985; Li and Mei,
1994), while α2 receptor stimulation has strong beneficial effects (Arnsten, 1997; Arnsten et
57
al., 1988; Arnsten and Goldman-Rakic, 1985; Franowicz and Arnsten, 1998; Mao et al., 1999;
Rämä et al., 1996; Tanila et al., 1996).
In summary, despite producing mainly one neurotransmitter, NA, LC neurons have
diverse inputs and targets, with a modular organization of their connections, diverse
electrophysiological properties, such as excitability, and diverse effects at the neuronal level
depending on the target area. In the following parts, I quickly review frequently used proxies
for LC activity, and then focus on functions of LC activity and NA release relevant to the
hypothesis of an implication in effort processing. I start with brief reviews of accounts of the
implication of the LC/NA system in resource mobilization for costly cognitive and physical
processes, starting with two global processes associated with overall increases in cognitive
and physical expensed resources: arousal and stress. I then review evidence for the
implication of the LC/NA system in attention. I finish with more detailed accounts of the
impact of LC activity on goal-directed functions requiring resource mobilization, i.e. effort
(cognitive control or physical effort), with a special focus on decision-making. Finally, I review
different mechanistic accounts for the effects of LC activity on target areas at the neuronal
and network level that could explain the behavioral effects.
2. Assessing LC activity: proxies for LC activity
a) P3 ERP
The P3 is a positive-going ERP (see Figure 18 for details). Although first found in humans
(and mostly used in humans), the P3 component can be found in a large variety of species:
monkeys, cats and rats (Arthur and Starr, 1984; Jodo et al., 1995; O’Connor and Starr, 1985).
P3 responses were shown to correlate with the activity of a large number of cortical and
subcortical areas, for example the PFC (Kirino et al., 2000; McCarthy et al., 1997), the parietal
cortex (Smith et al., 1990; McCarthy et al., 1997), the temporal lobe (McCarthy et al., 1989),
and the thalamus (Yingling and Hosobuchi, 1984). Overall, studies on the brain areas
responsible for the P3 component have pointed towards a relation with the temporal parietal
junction, the medial temporal complex and the lateral PFC (Soltani and Knight, 2000).
Considering the multiplicity and wide spatial distribution of brain areas contributing to
the P3 component, several teams have proposed that it might in fact be caused by the activity
of a neuromodulatory system sending projections throughout the brain, such as the LC
58
(Nieuwenhuis, 2011; Nieuwenhuis et al., 2011, 2005; Pineda et al., 1989). Indeed, the areas
whose activity has been linked to the P3 component happen to receive strong projections
from the LC. Moreover, P3 magnitude is affected by the temporal frequency of stimuli, with
greater magnitude for highly infrequent and therefore surprising stimuli (Polich, 2007;
Verleger et al., 1994). Such a pattern is reminiscent of the characteristic activation of LC
neurons by surprising and salient stimuli (Bouret and Richmond, 2009; Bouret and Sara, 2004;
Grant et al., 1988; Vankov et al., 1995), and the timing of the P3 responses would be
compatible with being caused by that LC response (Nieuwenhuis et al., 2005). Additionally, P3
responses are sensitive to attention, and strong proof for the implication of the LC/NA system
have been provided by a large body of literature, reviewed in part I.B.3.c).
Besides these indirect correlative accounts of the relationship between LC activity and
the P3 component, Swick and colleagues (1994) performed simultaneous recordings of LC
activity and P3-like potentials in monkeys, showing directly the correlation between LC bursts
of activity and the P3 component. Moreover, in order to establish the causality of the
relationship between LC activity and P3 responses, Pineda and colleagues (1989) recorded P3-
like potentials in monkeys after bilateral LC lesions. They report a decrease in the magnitude
of the P3-like component, with the decrease in magnitude correlating with the extent of the
lesions. De Rover and colleagues (2015) showed that blocking β adrenergic receptors with
propranolol modulated the amplitude of the P3 component in response to stimuli, in a
direction dependent of the participants’ initial level of arousal. More recently, Vazey and
colleagues (2018) used optogenetics to stimulate LC neurons, and showed that bursts of LC
spikes, but not low frequency discharge, elicited P3-like responses in the EEG (Figure 21).
Vazey et al., 2018, Figure 2B&D
Figure 21: ERPs evoked by LC optogenetic stimulation (adapted from Vazey et al., 2018).
Following laser simulation mimicking bursts of LC activity (pattern shown on the right side),
the ERP signal showed P3-like responses (330ms latency) (dark blue). It was not the case after
low frequency stimulation (cyan, pattern shown on the right side). The x-axis represents time
from optic stimulation, and the y-axis represents the recorded EEG signal.
59
Altogether, these studies indicate a correlation between LC activity and the P3. Note
however that this relationship lacks temporal precision and specificity, as the activity of many
other areas also correlates with these components.
b) Pupil dilation
Considering the difficulty of accessing the anatomical position of the LC and its relative
invisibility in regular MRIs, numerous teams routinely make the choice to approximate LC
activity with several physiological easily measured proxies, although these are less specific and
less precise (both regarding the brain areas correlating and regarding time dynamics). The
main proxy is pupil dilation, of particular interest for that manuscript considering the positive
relationship between pupil dilation and physical and mental effort (see part I.A.1.c)(3) and
part I.A.2.g)(2), respectively).
Several studies using electrophysiology (Joshi et al., 2016; Varazzani et al., 2015) or
functional magnetic resonance imaging (fMRI, Murphy et al., 2014) to measure LC activity
have reported positive correlations between LC activity and pupil dilation in different types of
tasks. The studies from Murphy and colleagues (2014) and Joshi and colleagues (2016)
however brought to light the lack of specificity of that measure since both reported
correlations between neuronal activity and pupil diameter in additional areas, including other
parts of the brainstem. A few pharmacological studies explored the causal relationship
between LC activity and NA availability in the brain and pupil diameter. Phillips and colleagues
(2000) used an α2 agonist (clonidine) to decrease the activity of LC neurons, or an α2
antagonist (yohimbine) to increase the activity of LC neurons, in healthy humans, and found
that pupil diameter decreased under clonidine and increased under yohimbine. Additionally,
patients suffering from a deficiency of the dopamine-β-hydroxylase (DβH), a critical enzyme
for NA synthesis, and therefore lacking brain NA, were shown to have abnormally small or
even absent pupil responses to task events (Jepma et al., 2011).
60
(A) Joshi et al., 2016, Figure 3C left panel (B)
(C) Murphy et al., 2014, Figure 2A (D) Murphy et al., 2014, Figure 2B
Figure 22: Positive correlation between pupil diameter and LC activity
(A,B) Positive correlation between pupil diameter (average over trial, z-scored, x-axis) and LC
rate (extracellular recordings, rate, y-axis), in example session. The lines represent the linear
fit of the data.
(A) Adapted from Joshi et al., 2016.
(B) Data pooled by bins for visual display (adapted from Varazzani et al., 2015).
(C,D) Positive correlation between LC blood-oxygen-level dependent (BOLD) signal and pupil
dilation.
(C) Whole-brain map showing brain regions whose unsmoothed BOLD signal was correlated
with pupil diameter at rest or during an oddball task, and overlap between the two.
(D) Areas of the pons with effect of pupil dilation on BOLD signal (colors indicate size of F), with
atlas-derived location of the LC highlighted. LC BOLD activity correlated with pupil dilation.
Note however that the areas correlating to pupil diameter in the brainstem were not restricted
to the LC (adapted from Murphy et al., 2014).
To summarize, two non-invasive proxies are quite frequently used as proxies for LC
activity, despite their lack of temporal precision and mediocre specificity. Note also that
studies looking for correlations between these two proxies have provided inconsistent results
(Kamp and Donchin, 2015; LoTemplio et al., 2021; Murphy et al., 2011).
61
3. LC activity and enhancement of processes requiring
resource mobilization
The hypothesis of the implication of the LC in effort stems from the various accounts
of its implication in processes involving global increases in cognitive and physical expensed
resources via global increases in cerebral and motor activity: arousal, stress and attention. I
will review the role of the LC/NA system in these three processes in the next parts.
a) LC activity and arousal
Arousal, the physiological and psychological state of alertness or awareness,
fluctuates, with two major states: sleep and wake. Within wake, arousal can vary from
drowsiness to high-arousal and stress. LC neurons are virtually silent during rapid eye
movement (REM) sleep, have a low rate during slow wave sleep, a moderate baseline rate
with bursts of activation for salient stimuli during stress-free waking, and a high baseline rate
during stressful events (Aston-Jones and Bloom, 1981). Arousal has been repeatedly shown to
be promoted by LC activity and NA discharge in the brain (Aston-Jones, 2005; Carter et al.,
2010; Foote et al., 1983; Jouvet, 1969; Saper et al., 2005; Sara, 2009; reviewed in Berridge and
Waterhouse, 2003). LC neurons are virtually silent during sleep (see Figure 47B in
experimental results for example, and changes in firing rate predict sleep-wake transitions.
Moreover, LC neurons respond to salient events, and these responses prolong waking periods
(Aston-Jones and Bloom, 1981; Foote et al., 1980; Hobson et al., 1975).
One widely used application of such impact of NA on arousal levels consists in
administrating α2-agonist drugs (clonidine, medetomidine, xylazine…) as sedatives for general
anesthesia (De Sarro et al., 1987; Gatti et al., 1988). Indeed, such drugs act on pre-synaptic
regulatory receptors of LC neurons, thus inhibiting NA neurotransmission and LC activity, as
sedatives for general anesthesia (De Sarro et al., 1987; Gatti et al., 1988). Coherently, blocking
the adrenergic receptors responsible for a large part of the post-synaptic excitatory effects,
i.e. α1 and β receptors, was shown to have sedative effects (although such pharmacological
methods are not routinely used for anesthesia, Berridge and España, 2000).
Considering the relatively low specificity and large diffusion of drugs targeting
adrenergic receptors, later accounts of the sedative effect of LC activity inhibition were made
62
using more precise techniques such as local infusions of drugs in the vicinity of the LC while
recording neuronal activity (Adams and Foote, 1988; Berridge and Foote, 1991). Conversely,
administration of NA itself directly into brain ventricles or the forebrain (Flicker and Geyer,
1982; Segal and Mandell, 1970) promotes wakefulness. Additionally, activation of the LC was
shown to promote forebrain activation (assessed via EEG, in anesthetized animals, Berridge
and Foote, 1991), and the NA-dependence of this effect was assessed via pretreatment with
noradrenergic inhibitors (α2-agonist and β-antagonist). Much recently, Carter and colleagues
(2010) used optogenetic techniques, that allow specifically targeting neuronal populations
such as NA producing neurons, to demonstrate the association between LC firing, cortical
activity, sleep-wake transitions, and arousal in general. Practically, they inhibited or
stimulated LC firing and observed that inhibition of LC activity promoted wake-sleep
transitions, without increasing the duration of sleeping periods, and that stimulating LC
activity caused sleep to wake transitions (Figure 23), thus providing precise and selective
evidence for a causal implication of LC activity in promoting waking states.
Overall, these accounts of the implication of LC activity in arousal provides a first
indication that the LC/NA system is involved in mobilizing resources for behavior, provided
that being awake requires more resources than being asleep (at least metabolic resources).
(A) Carter et al., 2010, Figure 3A (B) Carter et al., 2010, Figure 4A
Figure 23 : Optogenetic inhibition of LC activity promotes sleep while stimulation promotes
wake (Adapted from Carter et al., 2010).
(A) Optogenetic inhibition of LC activity (eNpHR-eYFP) reduces wake duration in proportion of
total time compared to control group (eYFP) and increases non REM sleep duration in
proportion of total time.
(B) Optogenetic stimulation of LC activity at 3Hz (ChR2-eYFP) increases wake duration in
proportion of total time and decreases sleep duration in proportion of total time, compared to
baseline (no stimulation) and control group (eYFP).
Abbreviations: eNpHR: halorhodopsin, yellow light sensitive chloride pump; ChR2:
channelrhodopsin-2, blue-light sensitive cation channel; eYFP: yellow fluorescent protein,
marker and control; REM: rapid eye movement sleep; NREM: non-rapid eye movement sleep.
63
*: p<0.05; **: p<0.01; ***: p<0.001.
b) LC and stress-related responses
Stress-related responses regroup a larger range of behavioral responses such as fear,
anxiety and avoidance. Stress promotes arousal, vigilance and attention, rendering the
cognitive resources necessary for an appropriate response to the stressing event available4. In
an effort to further explore the relationship between LC activity and resource mobilization for
costly processes, this part will review the litterature regarding the relationship between LC
activity and stress responses.
The LC is a central part of the ventral ‘fight-or-flight’ response system (Koob, 1999;
Sved et al., 2002), with strong reciprocal connections with the different regions of the
amygdala (Bouret et al., 2003; Buffalari and Grace, 2007; Van Bockstaele et al., 2001), a key
region for stress responses (LeDoux, 2000; McGaugh, 2004) and stress-related disorders
(Rauch et al., 2000; Shekhar et al., 1999). NA has been shown to be involved in the modulation
of responses to various types of stressors: social stress (Bingham et al., 2011), pain
(footshocks, Cassens et al., 1981), predator stress (Curtis et al., 2012) and homeostatic stress
(hemodynamic stress, Valentino et al., 1991). Several stress related disorders have been
associated with increased discharge or decreased sensitivity to inhibition of LC neurons (Koob,
1999; Olson et al., 2011; Raskind et al., 2013; Valentino and Van Bockstaele, 2008), and
therapeutic strategies for such conditions have focused on regulating noradrenergic effects.
For example, post traumatic stress disorders (PTSD), often encountered in people exposed to
prolonged stressful situations (e.g. soldiers during a war, abuse victims) have been successfully
treated with α1 adrenergic receptor antagonists such as prazosin (Raskind et al., 2013).
LC neurons respond strongly to stressors. Different studies have reported different
neuronal effects: increases of baseline rate of discharge (Curtis et al., 2012, 1997; Valentino
and Foote, 1988) or increases in rate evoked by salient stimuli and evoked response to
baseline (signal-to-noise) ratio (Bingham et al., 2011, see Figure 24A). Note that this
discrepancy could be due to the fact that, in rats, individuals in different age categories seem
to show different effects (Bingham et al., 2011).
4 Note that excessive stress tends to impair cognitive functioning, and at least some part of such impairment is
dependent on α1 adrenergic receptors (Birnbaum et al., 2004).
64
LC activity modifications in response to stressors were shown to be associated with
cortical EEG activation (Curtis et al., 1997). Moreover, it is necessary and sufficient for the
onset of anxiety behavior, although the opposite is not true, as inhibiting LC activity does not
produce anxiolytic effects (McCall et al., 2015). Additionally, stress-related LC activity
modifications were shown to be mediated by corticotropin-release hormone (CRH) secretion
in LC, most likely from axons emerging from the amygdala (McCall et al., 2015; Reyes et al.,
2008). Indeed, CRH microinfusions in the LC (Curtis et al., 1997; Page and Abercrombie, 1999),
as well as optogenetic stimulation of CRH-releasing fibers from the central amygdala (McCall
et al., 2015), produce robust increases in firing rate, and CRH antagonists prevent the increase
in rate of LC neurons in response to stress (Curtis et al., 2012; Valentino et al., 1991).
(A) Bingham et al., Figure 6A (B) Curtis et al., 1997, Figure 1A, bottom panel
Figure 24 : LC activity is modulated by stress, via the action of CRH.
(A) In rats, following social stress, LC neurons showed increased responses to salient stimuli
(grey) compared to control group not exposed to stress (black) (Adapted from Bingham et al.,
2011).
(B) In halothane anesthetized rats, after CRH microinfusion (30ng) in the LC, LC rate increased
immediately and remained higher than before infusion for 30 minutes (adapted from Curtis et
al., 1997).
In summary, an extensive body of literature points towards a preponderant role of the
LC in stress-related responses. For behavior to be adaptive, detecting stressful elements in the
environment has to trigger costly processes, and the fact that the LC activation is involved in
stress-related responses provides additional elements in favor of a role of the LC in mobilizing
energy for costly cognitive and physical processes.
c) LC activity and attention
First accounts of the role of the LC/NA system in attention were made in the 1980s
based on pharmacological and clinical studies (Clark et al., 1987; Robbins, 1984; Sara, 1985).
For example, clinically, selective NA reuptake inhibitors, such as atomoxetine, are efficient for
65
relieving the attention-related symptoms of patients suffering from ADHD (reviewed in
Clemow and Bushe, 2015).
Petersen and Posner (2012) described attention as composed of three sub-processes:
alerting, orienting and attentional executive control. In tasks requiring responding to certain
stimuli, if these stimuli are cued by a warning signal, RTs are shorter. This alerting effect
decreases in case of administration of α2 agonist drugs such as clonidine and guanfacine,
(Coull et al., 2001; Witte and Marrocco, 1997), suggesting that the LC/NA system may be
causally involved in attentional alerting. In coherence with that idea, individual LC neurons
produce a characteristic activation-inhibition pattern of response to novel salient stimuli, with
habituation if the stimulus is repeated a certain number of times (Bouret and Richmond, 2009;
Bouret and Sara, 2004; Grant et al., 1988; Vankov et al., 1995, Figure 20).
Attention orienting refers to the ability to focus attention at a particular point in time
and/or space. For instance, tasks used to test spatial orienting of attention consist in asking
subjects to report detecting a target after presenting either a valid (providing accurate
information on future target position) or invalid (inaccurate information) distractor. The
orienting effect consists in the observation of shorter RTs in the presence of valid distractors.
Several studies have reported an impact of pharmacological modifications of NA
neurotransmission on attentional orienting. Specifically, increases in brain NA with reuptake
inhibitor atomoxetine or optogenetic activation of LC neurons caused improvements of
orienting in the majority of subjects (Bari et al., 2020; Reynaud et al., 2019), while decreases
in LC activity in response to α2 agonist (clonidine) administration or optogenetic inhibition
caused deterioration of orienting (Bari et al., 2020; Clark et al., 1987; Coull et al., 2001). Note
however that this effect of NA on orienting is probably context dependent, as it was only
reported in highly predictive contexts (high proportion of valid targets) but was not found
otherwise (Witte and Marrocco, 1997).
Overall, several studies converge towards an implication of the LC/NA system in
attention, a process allocating resources to monitoring external stimuli. In later parts, I will
review a potential mechanism through which LC activity may promote attention: increased
signal-to-noise ratio in brain areas involved in sensory processing (see part I.C.2.a) for details).
This potential mechanism resonates with the finding from Reynaud and colleagues (2019) that
66
the effect of atomoxetine on spatial orienting is in part due to an enhancement of the sensory
information accumulation process.
4. LC and execution of costly goal-directed actions
In the previous parts, I reviewed evidence for the implication of the LC/NA system in
various processes involving resource mobilization: wakefulness, stress and attention. These
three however constitute cognitive states and are not goal-directed. In the following
paragraph, I review evidence for the implication of the LC/NA system in executing costly
actions, whether they be costly due to mental and/or physical effort requirements. Note that
I will focus specifically on the role of the LC in the exertion of cognitive control for decision-
making in a following part (I.B.5).
In a previous part of this manuscript, I reviewed extensively the positive relationship
between mental effort and pupil diameter, and between mental effort and the P3 component
of ERPs (see part I.A.1.c)(3). Since pupil diameter (Joshi et al., 2016; Murphy et al., 2014;
Varazzani et al., 2015) and the P3 (Swick et al., 1994; Vazey et al., 2018) are proxies for LC
activity (see part I.B.2), indirectly, LC activity could increase during mental effort exertion.
Evidence for a causal role of LC activity in improving cognitive functioning and in particular
mental effort was provided by studies using pharmacological modifications of LC activity
(reviewed by Robbins and Arnsten, 2009). Such studies were done in patients suffering from
neuropsychiatric diseases reducing cognitive control capacity, and in healthy humans and
animals, with numerous studies focusing on different components of mental effort separately:
working memory, inhibitory control and task-switching.
Atomoxetine, a reuptake inhibitor of NA, increases NA levels throughout the brain and
especially in frontal cortices. It is routinely used in the treatment of ADHD. This disorder is in
particular associated with impulsivity, and difficulties with tasks requiring inhibition, working
memory and task-switching (DSM-5, American Psychiatric Association and American
Psychiatric Association, 2013). This symptomatology is believed to be due to an impairment
of the capacity for cognitive control exertion (e.g. Arnsten and Rubia, 2012; Geurts et al., 2004;
Pennington and Ozonoff, 1996; Sergeant et al., 2002). The alleviation of ADHD symptoms by
atomoxetine indirectly suggests that NA brain levels may be critical for efficient cognitive
control exertion (Chamberlain et al., 2007).
67
Additionally, atomoxetine was shown to improve response inhibition, causing a
decrease in impulsive behavior in rats (e.g. Arnsten, 2006; Fernando et al., 2012; Navarra et
al., 2008; Robinson, 2012; Robinson et al., 2008), healthy humans (Chamberlain et al., 2009)
and patients suffering from a large range of neuropsychiatric conditions (reviewed by Pattij
and Vanderschuren, 2008). Atomoxetine as well as another reuptake inhibitor of NA,
desipramine, were shown to improve stop signal reaction times (SSRTs, an estimation of the
time taken to attend to and respond to a stop signal following a go signal) in humans
(respectively: Chamberlain et al., 2006a; Overtoom et al., 2003) and rats (atomoxetine:
Robinson et al., 2008). Atomoxetine administration in the context of ADHD was also shown to
improve Stroop task performance (Faraone et al., 2005). Although atomoxetine is not
perfectly specific of NA reuptake, and also tends to increase DA levels in the brain (Bymaster
et al., 2002), it was shown that the effects of atomoxetine on inhibitory control were not
reproduced by selective DA reuptake inhibitors (Bari et al., 2009), reinforcing the idea of a
critical role of NA. Additionally, in Go/No-Go tasks, α2 receptor blockade by yohimbine in
primates was shown to impair performance, suggesting that increases in NA availability may
act on response inhibition by stimulating these receptors (Ma et al., 2003).
Regarding the implication of the LC/NA system in inhibitory control, clear evidence for
a causal role of LC activity was provided by an optogenetics study by Bari and colleagues (2020,
Figure 25). They showed that optogenetic manipulations of LC activity in a task requiring
response inhibition (Figure 25A) modified performance: increasing LC activity increased
performance Figure 25C&D) while decreasing LC activity decreased performance (Figure
25E&F). Additionally, mice were trained on a modified Posner task (Figure 25B), in which
distractors either predicted the position of a target (valid distractor), provided no information
(neutral distractor) or indicated a position different from the future position of a target. In
that task, the influence of distractors on RT decreased in case of LC stimulation (Figure 25G&H)
and increased in case of LC inhibition (Figure 25I&J).
Finally, using Calcium imaging, the authors report finding, first, that responses with
lesser effects of distractors (less effect of distractor validity on RT and less errors) and better
inhibitory control (correct versus premature responses) were preceded by higher
spontaneous LC activity (Figure 25K-P) and stronger responses to distractor onset (Figure
68
25M) than responses with strong effects of distractors and insufficient inhibitory control
exertion. Second, responses to target onset were greater before fast responses (Figure 25Q-S).
Altogether, these results would indicate that the baseline rate of LC neurons as well as
the strength of their pre-action (post-target onset) responses are causally related to the
strength of the cognitive control exerted throughout a trial, at least in the context of inhibitory
control exertion.
(A) Bari et al., 2020, Figure 2A (B) Bari et al., 2020, Figure 3A
(C) Bari et al., 2020, Figure 2E (D) Bari et al., 2020, Figure 2F
(E) Bari et al., 2020, Figure 2I (F) Bari et al., 2020, Figure 2J
(G) Bari et al., 2020, Figure 3D (H) Bari et al., 2020, Figure 3H
(I) Bari et al., 2020, Figure 3F (J) Bari et al., 2020, Figure 3J
(K) Bari et al., 2020, Figure 4G (L) Bari et al., 2020, Figure 4H
(M) Bari et al., 2020, Figure 4I (N) Bari et al., 2020, Figure 4J
(O) Bari et al., 2020, Figure 4K (P) Bari et al., 2020, Figure 4L
(Q) Bari et al., 2020, Figure 4M (R) Bari et al., 2020, Figure 4N
(S) Bari et al., 2020, Figure 4O
Figure 25: Optogenetic manipulations and Calcium imaging of LC activity demonstrate a causal
role of the LC/NA system in inhibitory control exertion (adapted from Bari et al., 2020)
(A) Schematic representation of the two-choice task. Mice had to wait for the appearance of a
target in one of two locations during a delay and poke target location to receive a food reward.
Inhibitory control was required to avoid premature responses and sustained attention was
required to choose the correct target location. LC activity was modified during the delay.
(B) Modified Posner task. After a certain delay, a distractor could (cued condition) appear or
not appear (no-cue condition). Then the target appeared and the mice had to poke its location.
The distractor could either provide neutral, valid or invalid information about target location.
The task required response inhibition to refrain from premature responses and to avoid
incorrect responses due to distractors. LC activity was modified during the delay.
(C-F) Two-choice task: Effects of optogenetic (C,D) activation or (E,F) inhibition of LC (blue or
green bars, respectively) compared to control (grey) on (C,E) performance (percentage of
69
correct responses) and (D,F) premature responses. LC activation increased performance and
decreased premature responses, while LC inhibition produced the opposite pattern of results,
suggesting an enhancement of selective attention and inhibitory control.
(G-J) Modified Posner task: Effects of optogenetic (G,H) activation or (I,J) inhibition of LC
activity (filled blue or green dots respectively) compared to control (unfilled dots) on (G,I)
performance validity effect (percentage correct in valid - invalid conditions) and (H,J) RT
validity effect (RT valid - RT invalid). LC activation reduced validity effects and LC inhibition
enhanced validity effect, suggesting a causal role of LC activity in inhibitory control exertion.
(K-S) Calcium imaging of LC activity in the modified Posner task, during (K,L,N,O,P) pre-
distractor period (baseline), (M) distractor presentation, and (Q,R,S) target presentation.
Correct responses were preceded by higher baseline activity (K,L) compared to (K) incorrect
and (L) premature responses, and by (M) higher activity during distractor presentation
compared to premature responses. (N-P) Higher baseline activity and (Q-S) higher activity
during target presentation was associated to lesser validity effect on RT. (O,P) Increased
baseline and (R,S) post-target onset activity preceded (O,R) slow valid and (P,S) fast invalid
trials, compared to (O,R) fast valid and (P,S) slow invalid trials, respectively. There was no
difference in (N) baseline activity or (Q) target-evoked activity between fast and slow trials
with neutral distractors.
*: p<0.05; **: p<0.01; ***: p<0.001; n.s.: non significant; TW: time window, F/F: difference
between recorded signal (fiber photometry) and mean signal, divided by mean signal
(normalized), Corr: correct; incorr: incorrect ; prem: premature.
Regarding working memory, an early study using intra PFC injections of an α2
antagonist (yohimbine), and α1 antagonist (prazosin) or a β antagonist (propranolol)
demonstrated the critical role of α2 adrenergic receptors in working memory, since only α2
adrenergic receptor antagonism had effects on performance in a spatial working memory task
(Li and Mei, 1994). This finding was confirmed by various other pharmacological studies (e.g.
Gamo et al., 2010; Li et al., 1999; Mao et al., 1999; Tzavara et al., 2006; Wang et al., 2007) as
well as LC lesion studies (Coradazzi et al., 2016). Overall, β receptor blockade (propranolol) as
well as α2 receptor stimulation (clonidine) impaired working memory while NA reuptake
70
inhibitors enhanced working memory (reviewed by Chamberlain et al., 2006b; Ellis and
Nathan, 2001).
Based on their pattern of response to changes in rules, LC neurons could participate to
the learning of new task contingencies, and therefore to set-shifting performance. LC
responses to new task-relevant stimuli during learning appear early on, after only a few trials,
long before behavior adjusts to the new rules (Bouret and Sara, 2005). Early causal accounts
of the relationship between NA and set-shifting were provided by Devauges and Sara (1990),
through pharmacological enhancements of LC activity, and McGaughy and colleagues (2008),
through lesions of noradrenergic projections to the PFC. Additionally, Lapiz and Morilak (2006)
showed that α2 antagonist (atipamezole) administration improved set-shifting performance
by increasing NA availability for α1 receptors (effects of atipamezole were blocked by α1
antagonist administration). Similarly, increasing NA availability with acute or prolonged
administration of desipramine, a NA uptake inhibitor, in rats, increased set-shifting
performance (Lapiz et al., 2007), while selectively decreasing cortical NA availability through
lesions of the LC cells projecting there altered set-shifting performance (McGaughy et al.,
2008; Tait et al., 2007).
Besides facilitating adaptation to new task rules when the task imposes it, the LC/NA
system was shown to have a critical implication in switching from exploration to exploitation.
Indeed, in an exploration-exploitation tradeoff gambling task, Jepma and Nieuwenhuis (2011)
found that pupil dilation, which they used as a proxy for LC activity, increased more before
exploratory choices than exploitative choices, suggesting that brief activations of LC neurons
could lead to disengagement from exploration (Figure 26A). Moreover, individual pupil
baseline predicted individual tendencies for exploration, suggesting a relationship between LC
baseline activity and exploration (Figure 26B).
(A) Jepma ans Nieuwenhuis, 2011, Figure 2A (B) Jepma ans Nieuwenhuis, 2011, Figure 4A
Figure 26: Pupil diameter increased more before exploration than exploitation, and individual
pupil diameter baseline was positively correlated with individual tendency to explore (adapted
from Jepma and Nieuwenhuis, 2011).
71
(A) Time course of average pupil diameter aligned on trial-by-trial response (key press). Before
and after key press, pupil diameter was larger in exploratory trials.
(B) Positive correlation between average pupil diameter in mm (x-axis) and proportion of
exploratory choices (y-axis). Subjects exploring more had a larger average pupil diameter.
Gilzenrat and colleagues (2010) obtained similar results, with some complements. In
three distinct experiments, they demonstrated that exploration periods (task disengagement)
were associated with high pupil baseline while exploitation periods (task engagement)
corresponded to low pupil baseline with strong evoked pupil responses to task events.
Although pupil diameter is merely an imperfect proxy for LC activity, these results could
suggest that LC neurons spike at a high baseline rate during exploration and with a low
baseline but strong responses to task events during exploitation.
Providing proof of the causal link between LC activity and exploration/exploitation
tradeoff, increasing LC activity with DREADDs (designer receptor exclusively activated by
designer drugs) caused disengagement from exploration tasks and increased choice variability
(Kane et al., 2017, Figure 27A&B). Conversely, decreasing NA with clonidine (an α2 agonist
decreasing LC activity) led to decreased choice variability (Jahn et al., 2018, Figure 27C),
indicating a role of the LC in promoting behavioral variability and exploration.
(A) Kane et al., 2017, Figure 4, top left panel
(B) Kane et al., 2017, Figure 4, bottom left panel
(C)
P(c
ho
osi
ng
alte
rnat
ive
op
tio
n
Value of alternative option –
value of current option
72
Figure 27: Increasing LC activity caused increased exploration and decision variability, while
decreasing LC activity decreased choice variability.
(A) In a patch-foraging task, regardless of initial patch value i.e. initially available reward (x-
axis), increasing LC activity via DREADDs stimulation with clozapine-N-oxide (CNO) (blue and
green curves) decreased the time (number of trials) spent in the patch (y-axis) compared to
placebo (red) (figure adapted from Kane et al., 2017).
(B,C) Increasing LC activity increased choice variability.
(B) In a patch-foraging task, increasing LC activity caused the relationship between the
expected value of leaving the current patch (x-axis) and the probability of leaving the patch (y-
axis) to become shallower under CNO treatment (blue and green) compared to placebo (red).
(C) In a value-based choice task, decreasing LC activity by clonidine administration (orange)
caused increased choice variability compared to placebo (brown), materialized by a steeper
curve of the relationship between the value difference of the options (x-axis) and the
probability to choose the alternative option (y-axis) (figure adapted from Jahn et al., 2018).
More recently, Dubois and colleagues (2021) showed that, in an exploration-
exploitation trade-off task, administration of β receptor antagonist propranolol cause a
reduction of random exploration, i.e. of noise in choices. Note that propranolol could have
acted by increasing availability of NA (since it did not bind β receptors) and therefore increased
α2 receptor stimulation.
Considering the hypotheses regarding the aversive nature of cognitive control being
related to a regulation of the exploration-exploitation tradeoff, the involvement of the LC in
switching from exploitation to exploration could reflect an involvement of the LC in regulating
cognitive control allocation. This role could seem incompatible with its probable involvement
in promoting control engagement in working memory and response inhibition. However, set-
shifting and behavioral flexibility seem to depend on different receptors, with different
affinities for NA, than working memory and inhibitory control. In coherence with that,
atomoxetine doses that improve response inhibition (stop-signal performance in rats) appear
to be detrimental for set-shifting (Newman et al., 2008). Consequently, low doses of NA (low
firing rates of the LC) could be responsible for control exertion enhancement, while high doses
73
of NA (high firing rates of the LC) could be required for shifting from investing control in a
specific task (exploitation) to exploring other options. In other words, the relationship
between LC activity, NA levels, and cognitive control enhancement probably follows an
inverted U shaped curve, in which increasing NA facilitates control exertion but only up to a
certain point.
Regarding physical effort exertion, LC neurons have been shown to encode the exerted
force during force-discounting tasks, at the time when the force is exerted (Varazzani et al.,
2015). Additionally, pupil diameter has been shown to increase during physical effort exertion
in humans, and to correlate both with the objective and subjective perception of effort (Zénon
et al., 2014, see also Figure 3). These results would tend to indicate an implication of LC activity
in effortful physical actions, yet this increase in LC activity and pupil diameter could merely
reflect autonomic activation due to muscle contraction. However, pharmacological evidence
have complemented these results by providing evidence for a causal relationship between LC
activity and force production. Two studies, consisting in administrating an α2 agonist
decreasing LC activity, clonidine, to monkeys executing tasks requiring force exertion,
reported decreased force exertion (Borderies et al., 2020; Jahn et al., 2018). Interestingly, in
Borderies et al. (2020, in annex), we reported that this effect of clonidine was not the result
of a decreased capacity to exert force, at the muscular level, since the muscular contractility
index was not affected by the treatment. This indicates that decreased LC activity may cause
physical effort to be perceived as more aversive.
(A) (B) (C)
Figure 28: LC neurons encoded the exerted force during force exertion, and decreasing brain
NA decreased exerted force.
74
(A) Positive correlation between the firing of LC neurons around action onset (x-axis) and the
maximum force exerted by the monkeys (y-axis) during a one-option force-discounting task
(figure adapted from Varazzani et al., 2015).
(B) Exerted force (y-axis) in a reward-schedule task following placebo or clonidine
administration at different doses (x-axis). Clonidine administration decreased the exerted force
(figure adapted from Jahn et al., 2018).
(C) Maximum exerted force in percentage of the maximum force that the monkey can exert (y-
axis) in a two-options force-discounting task, after placebo (grey) or clonidine administration
(brown), as a function of the requested force (y-axis). Dashed line represents the Force target
i.e. percentage of the maximum force that had to be exerted to get the Reward, by condition.
Clonidine administration decreased the exerted force (figure adapted from Borderies et al.,
2020).
Altogether, numerous studies pointed towards a causal role of LC activity in all three
subcomponents of cognitive control: response inhibition, updating working memory, and task
switching. Generalizing effects on set-shifting, high levels of LC activity lead to disengagement
from task and exploration. Additionally, LC activity appears to be critical for physical effort.
Considering that cognitive control and physical effort both constitute decision variables (i.e.
subjects try to avoid having to make efforts, see parts I.A.1.a) and I.A.2.d)(2)) and that
cognitive control is beneficial for decision-making (see part I.A.2.d)(1)), I will now specifically
review evidence for the implication of the LC in decision-making.
5. LC and decision-making
In the following part, I will split the decision-making process into two sub-processes:
committing to a choice (action selection) and triggering the chosen action. Both steps are
susceptible of requiring cognitive control investment. For instance, at the stage of action
selection, cognitive control is required to choose with as little biases as possible. When a costly
action has been chosen, triggering it requires exerting control to override an automatic
(default) response consisting in not doing anything (see parts I.A.1.a), I.A.2.d)(2) and I.A.1.a)
and I.A.2.d)(2)). Here, I review the evidence regarding the role of the LC/NA system in these
two steps.
75
a) Action selection
Several pharmacological studies and lesion studies, in which LC activity or the effects
of NA on LC targets were manipulated over quite long periods of time, have indirectly pointed
towards a potential causal role of LC activity in the action selection process. Motivational
salience refers to the regulation of the intensity of the behaviors facilitating the attainment of
certain goal, and in that constitutes a prerequisite of action selection. Following the finding
that exposure to salient stimuli increases NA in the pre-frontal cortex (PFC) and nucleus
accumbens (Ventura et al., 2008, 2007), Ventura and colleagues (2007) decreased NA levels in
the medial PFC (mPFC) by producing selective destruction of LC neurons targeting the mPFC.
They found that mice with lower NA levels in the mPFC exhibited impaired motivational
salience.
Regarding action selection per se, pupil dilation studies provided indirect evidence for
the implication of the LC. Pupil dilation has been shown to increase with decisions (Beatty,
1982; de Gee et al., 2014; Einhäuser et al., 2008,2010; Gilzenrat et al., 2010; Lempert et al.,
2015; Nassar et al., 2012), especially when subjects respond against a bias, i.e. inhibit an
automatic response (de Gee et al., 2020, 2017, 2014). Considering the correlation between LC
activity and pupil diameter, and between cognitive control and pupil diameter, such findings
could indicate that the LC plays a role in allocating the control required to make decisions and
especially inhibit automatic responses. Additionally, in that very context, de Gee and
colleagues (2017) showed that the blood-oxygen level dependent (BOLD) fMRI signal in a
cluster of voxels including voxels compatible with the theoretical location of the LC was
negatively correlated with the strength of the decision biases (Figure 29A). This study further
advocates for the role of the LC in control exertion for decision-making and specifically action
selection, despite lacking spatial precision as well as not demonstrating a causality link.
76
(A) De Gee et al., 2017, Figure 2E left panel
(B) (C)
Figure 29: Effects of LC activity on action selection: increased LC activity, measured indirectly
through pupil dilation, was associated to decreased decision biases, while decreases in LC
activity by clonidine increased physical effort sensitivity.
(A) In a visual-detection task, the BOLD fMRI signal of a cluster of voxels corresponding to
brainstem neuro-modulatory nuclei including the LC (x-axis) was negatively correlated to the
strength of the bias in the evidence accumulation process of the decision (y-axis) (adapted from
de Gee et al., 2017).
(B) In a force-reward two-alternative choice task, decreasing brain NA with clonidine caused
an increase in the sensitivity of choices to the physical effort, materialized by a steeper slope
of the percentage of high force (HF) choices (y-axis) as a function of the absolute difference in
offered forces (|ΔForces|, x-axis) (figure adapted from Borderies et al., 2020).
(C) In the same task, clonidine did not affect the sensitivity to reward size, materialized by the
slope of the percentage of high reward (HR) choices (y-axis) as a function of the absolute
difference in offered rewards (|ΔRewards|, x-axis) (figure adapted from Borderies et al., 2020).
Complementing these findings, in Borderies et al., 2020 (in annex), we provided
evidence for a causal role of LC activity in physical effort-based action selection. We reported
that α2-agonist (clonidine) administration before a value-based decision-making task, with
options differing along the reward and physical effort dimensions, caused monkeys to be more
77
sensitive to the physical effort dimension while remaining equally sensitive to the reward
(Figure 29B&C). In other words, decreasing LC activity and brain NA levels influenced action
selection by making physical effort more aversive. Additionally, decreasing brain NA levels
caused monkeys to exert less force (Borderies et al., 2020; Jahn et al., 2018). We suggested
that decreasing NA increased the cost of effort, a computational variable representing the
amount of theoretical resources necessary for the execution of actions. If the cost of effort
increased, monkeys would theoretically avoid choosing high-effort offers more often, and this
would also have deleterious effects on action execution (force exertion). Thus, LC activity
appears to be critically but indirectly involved in the effects of physical effort on decisions.
Additionally, this finding could be interpreted in terms of effects on cognitive control
exertion for the decision process. Naturally, subjects have a bias towards avoiding high-effort
options. Consequently, choosing options associated with higher physical effort requires
exerting cognitive control, to override the default response. The decrease in high-effort
choices following decreases in LC activity could in fact indicate that LC activity is critical for
cognitive control exertion, at least for decisions. This idea also resonates with the negative
relationship between pupil dilation and decision biases and therefore putative decision
control exertion (de Gee et al., 2020, 2017, 2014).
Direct assessments of LC activity via in situ electrophysiological recordings, to our
knowledge, were never performed during value-based choice tasks with more than one offer
presented at a time (up to now, see experimental results of Study 2, part II.B). Consequently,
in one-option tasks, the choice that the subjects faced was between engaging in the trial, i.e.
trying to perform the requested action, and refusing the trial, i.e. not trying at all. Jahn and
colleagues (2020) report that, around cue onset in a one-option force-discounting task, LC
neurons differentiated between trials in which the monkeys would or would not engage, with
a higher spike count when the monkey would engage (Figure 30A). In the same vein, LC
neurons activated after a fixation break5 i.e. after disengagement from a trial (refusal to
engage), and the magnitude of this activation predicted engagement in the next trial (Figure
30B). Note however that this task did not allow differentiating value-based refusals from
putative automatic refusals of trials. Considering, however, that engagement was affected by
5 In this task, monkeys had to maintain gaze fixation onto the screen throughout trials, and therefore breaking fixation aborted trials.
78
task parameters in directions coherent with value-based refusals, I will interpret their results
under the hypothesis that monkeys performed a value-based analysis and cost-benefit
tradeoff to choose whether to engage in trials.
(A) (B) Engaged Not engaged Next trial: no engagement Engagement Cue Cue Fixation break Fixation break
Time from cue onset (s) Time from fixation break (s)
Figure 30: LC neurons encoded engagement in the current trial at the time of cue onset and in
the next trial around fixation breaks i.e. disengagement from trials (adapted from Jahn et al.,
2020).
(A) Raster plots of the activity of a representative LC unit aligned on cue onset, with trials split
between engaged trials (left, monkey will try squeezing the bar) and not engaged trials (right,
monkey will not try). The light brown line represents the spike density function. Around cue
onset in a force discounting-task, LC neurons positively encoded whether the monkey would
engage in the trial, i.e. not break fixation and try to perform the requested action (regardless
of the result of the action i.e. correct or incorrect).
(B) Raster plots of the activity of a representative LC neuron around fixation breaks. Trials are
split between fixation breaks followed by no engagement in the next trial (left) or followed by
engagement in the next trial (right). The light brown line represents the spike density function
and the grey dots represent cue onset. This neuron was activated around fixation breaks and
stronger activation predicted engagement in the next trial.
Choosing to engage in trials requiring force exertion can be considered as a form of
cognitive control exertion, since the default non-costly response would be refusing trials. In
79
refused trials, monkeys may have invested too little control and decided to forego the trial
even if this decision reduced reward rate, given that refused trials were repeated until
completion. The increased spike count of LC neurons around the presentation of cues when
monkeys would choose to engage (over and above the effects of task parameters) would be
compatible with a role of LC neurons in mobilizing cognitive control. Indeed, increased spike
count early after cue onset would have resulted in increased cognitive control exertion and
therefore higher probability to engage.
Selecting actions appropriately requires extensive information processes. Around the
time of presentation of information relevant for value-based decisions whether to engage in
trials, LC neurons exhibit an activation (Bouret et al., 2012; Bouret and Richmond, 2009; Jahn
et al., 2020; Varazzani et al., 2015). Several studies provided evidence of the encoding of the
amount of novel information carried by cues. Reward-schedule tasks consist in performing
series of presses to obtain a reward. The number of presses is variable, and along those series,
cues inform the monkeys of the number of remaining presses. Consequently, only the first cue
carries novel information regarding the length of the sequence, and the following ones only
duplicate that information. Notably, LC neurons have been shown to respond more strongly
(if not exclusively) to the cues signaling the first press of the sequence (Bouret et al., 2012;
Bouret and Richmond, 2009). In coherence with that, in a one-option force-discounting task,
Jahn and colleagues (2020) showed that, at the time of the presentation of the information
about the offer, LC neurons positively encoded the amount of information obtained. Indeed,
in case of refusal, trials were repeated, making a certain proportion of the cues uninformative
(providing redundant information), and LC neurons’ responses differentiated between
informative and uninformative cues (Figure 31).
80
(A) 1st cue Not 1st cue (B) New information No new information
Cue Cue Cue onset Cue onset
Time from cue onset (s)
Figure 31: LC neurons differentiated between cues carrying novel information and cues
carrying no novel information.
(A) Raster plots of the activity of an example LC unit in a reward-schedule task, around the
onset of the first cue of the sequence, carrying novel information about sequence length (left)
and other cues throughout the sequence (indicating the number of remaining presses, right),
carrying no additional information. Red and pink lines represent the spike density function. The
neuron exhibited no response to cue onset when the cue carried no additional information, and
activated in response to cue onset when the cue carried novel information (adapted from
Bouret et al., 2012).
(B) Raster plots of the activity of a representative LC neuron in a one-option force-discounting
task, contrasting trials with novel information associated to the cue (left) and trials with no
novel information carried by the cue (right) (adapted from Jahn et al., 2020). Raster plots show
a much weaker response of LC neurons to cue onset when the cue carries only redundant
information. The brown line represents the mean firing rate of the neuron across trials.
Moreover, LC neurons encode some of the information carried by the cues informing
subjects about the contingencies of trials, especially the offered reward (Bouret and
Richmond, 2015; Varazzani et al., 2015, Figure 32). In coherence with that, during economic
choices, pupil dilation has been shown to increase with expected value (Fiedler and Glöckner,
2012).
81
(A) (B)
Figure 32: LC neurons encoded some of the information provided by task cues on trial
contingencies, especially reward size.
(A) Raster plot of the activity of an example LC unit in a task consisting in pressing a grip for a
reward, around the onset of the cue indicating reward size. Trials are split by reward size,
darker blue indicates greater rewards. This unit encoded reward positively around cue onset
(adapted from Bouret and Richmond, 2015).
(B) Raster plots of the activity of an example LC neuron in a force-discounting task, around the
onset of the cue indicating the force-reward contingency. Trials are split along two dimensions:
offered reward (blue, horizontally) and requested force (red, vertically). Blue lines represent
the spike density function. This unit encoded reward size positively (adapted from Varazzani et
al., 2015).
b) Action triggering
LC neurons have been known for close to two decades to exhibit activation,
materialized by a few (sometimes even only one) spikes, slightly prior to action initiation
(Bouret et al., 2012; Bouret and Richmond, 2015, 2009; Bouret and Sara, 2004; Clayton et al.,
2004; Varazzani et al., 2015), suggesting a potential implication of LC activity in the action
triggering process. Additional evidence comes from the recording of LC activity in Go/No-Go
tasks. Kalwani and colleagues (2014), in a Go/No-Go task consisting in triggering responses to
82
a stimulus (saccade) and sometimes receiving the order to abort that response, found that LC
neurons activated around responses consisting in a motor response, both correct (Go trials)
and incorrect (No-Go trials). They conversely did not activate around responses consisting in
the absence of production of a motor response, even if that response was the correct one
(No-Go trials). This type of task is however not value-based, and I will from now on focus on
the role of LC activity in triggering costly actions, which is believed to require exerting cognitive
control.
In the same vein, Bouret and Richmond (2015) used a paradigm in which some trials
were active, i.e. the monkey needed to perform an action (bar release), followed by a feedback
and a reward, and some trials were passive, i.e. the reward was delivered without any action,
after a “feedback”. Noticeably, LC neurons activated prior to feedback onset in active trials
(i.e. prior to action onset) but not in passive trials (Figure 33). These patterns of activity of LC
neurons eliminates the possibility that pre-action LC activation corresponds to reward
expectation and suggests that LC neurons may be involved in the action triggering process.
In tasks involving responding to a cue serving as a go-signal as soon as a decision has
been reached, LC neurons only responded once between cue onset and action onset, and that
activation was better aligned on the onset of the action than on the cue (Bouret and Sara,
2004; Clayton et al., 2004). In tasks separating in time the processing of the information
provided by the cue, during a waiting period, and the response period, following a go-signal,
LC neurons responded both to the cue and before the action onset (Jahn et al., 2020; Varazzani
et al., 2015). This pre-action activation could be compatible with an implication of the LC in
the cognitive effort required to trigger costly actions.
This idea is especially reinforced by evidence in these tasks for a dissociation in time of
the action-selection process and action triggering process. Varazzani and colleagues (2015)
and Jahn and colleagues (Jahn et al., 2020) used a force-discounting task in which monkeys
were instructed to fixate their gaze on the location of the apparition of the cue. Fixation breaks
were thus considered as errors and led to the trial restarting. Interestingly, monkeys virtually
never broke fixation after the onset of the go signal, suggesting that the decisions whether to
engage in the trial based on the offered reward and requested force were made during the
waiting period of cue presentation. In that sense, this task probably allowed dissociating the
action selection process, during the waiting and cue presentation period, from the action
83
triggering process, after the go-signal. The two activations of LC neurons could be intervening
in the control exertion for these two steps separately.
(A) Clayton et al., 2004, Figure 2C (B) Clayton et al., 2004, Figure 3A
(C) (D)
Figure 33: In forced-choice discrimination tasks, LC activity was better aligned on response
onset than on stimulus onset. LC neurons were only active if an action was executed, regardless
of reward expectation.
(A) Raster plot (top) and peri-event time histograms (bottom) of the activity of an example LC
unit in the forced-choice discrimination task. LC activity is aligned on response onset (adapted
from Clayton et al., 2004).
(B) Population stimulus-locked peri-event time histograms with trials split between correct
and incorrect lever presses and omissions (no press). Bottom curves represent the density
function of RTs, split between correct and incorrect presses. LC neurons responded prior to the
onset of correct and incorrect responses (lever presses) but not in the case of omissions. Activity
was stronger prior to correct compared to incorrect responses (adapted from Clayton et al.,
2004).
(C) Raster plots (top) and peri-event time histograms (bottom) aligning the activity of an
example LC unit in an odor-discrimination task on either stimulus onset (CS+, left panel) or
response onset (Go resp, right side panel). LC activity is better aligned on response onset than
on stimulus onset (adapted from Bouret and Sara, 2004).
(D) Raster plot of the activity of an example LC unit around feedback onset in the task used by
Bouret and Richmond, 2015. The task consisted of a cue indicating reward size, displayed
Feedback
Active
Passive
84
during a waiting period, followed by either an active trial, in which monkeys had to release a
grip after a go-signal to get the reward, or a passive trial, in which the reward was delivered
after a certain amount of time. The reward was always preceded by a feedback, either when
the action was detected (active trials) or when the pre-defined delay had elapsed (passive
trials). On the plot, trials are split between active (black) and passive (grey). The activity of LC
neurons increased prior to the feedback in active but not in passive trials i.e. increased prior to
action onset.
If tasks conditions differ by their level of reward and/or costs (physical effort), during
the activation period prior to action onset, LC neurons have been shown to encode task
parameters in different directions and magnitudes compared to the modulations observed
around cue onset. For instance, in a reward schedule tasks in which some trials were rewarded
and some were not, an information given by the cue, LC neurons differentiated between
rewarded and unrewarded trials before action onset, with more spikes in the case of
unrewarded trials (Bouret et al., 2012, Figure 34A. This study additionally included the
measurement of the frequency of occurrence of a Pavlovian appetitive behavior, lipping,
known to signal the value of ongoing events (Bouret and Richmond, 2015, 2009). Interestingly,
lipping around action onset was more frequent in rewarded than non-rewarded trials,
indicating that monkeys perceived the action as more valuable, i.e. less costly when it led to a
reward. Consequently, triggering unrewarded actions probably required exerting more
cognitive control, and this was associated to increased LC activity.
Note that this pre-action negative encoding of reward was not seen in several other
studies (Bouret and Richmond, 2015; Varazzani et al., 2015, Figure 34B). It is interesting to
note that one of those studies included lipping behavior recording (Bouret and Richmond,
2015), and lipping probability of occurrence around action was not affected by reward either.
This suggests that the negative encoding of reward by LC neurons may indeed reflect the
subjective cost of triggering the action and therefore possibly the cognitive control invested,
as it is only present when rewards have an influence on behavior. Additionally, in a one-option
force-discounting task, Varazzani and colleagues (2015) reported a positive encoding of the
force level of the engaged action just prior to the action itself (Figure 34C). In the same line as
previous interpretations, this modulation by the requested force could indicate that LC
85
neurons encoded the cost of triggering the action, and possibly the invested cognitive control.
Note that I reanalyzed this data looking for additional evidence of this hypothesis (see
experimental results of study 1, part II.A).
(A) Rewarded Unrewarded (B)
Action Action
(C) Action
Figure 34: LC neurons encode task parameters just prior to action onset: reward negatively and
force positively.
(A) Raster plot of the activity of LC neurons around action onset in reward schedule task, with
trials split between rewarded (left) and unrewarded trials (right side). Blue lines represent the
spike density function. Prior to action onset, LC neurons activate and this activation is stronger
in unrewarded trials (adapted from Bouret and Richmond, 2012).
Firi
ng
rate
(H
z)
Firi
ng
rate
(H
z)
Firi
ng
rate
(H
z)
Firi
ng
rate
(H
z)
Firi
ng
rate
(H
z)
86
(B) Population activity of LC neurons recorded in a force-discounting task. Trials are separated
by required force level (effort 1,2,3) and offered Reward level. Pre-action activity and peri-
action activity are stronger when the action triggered involves a higher force level.
(C) Population activity of LC neurons recorded in a force-discounting task. Trials are separated
by required force level (effort 1,2,3). Pre-action activity and peri-action activity are stronger
when the action triggered involves a higher force level (Varazzani et al., 2015)
It is important to note, however, that the timing of the activity of LC neurons around
action onset is quite close to the action onset itself (roughly 250ms prior to action onset). Such
a timing is probably incompatible with an involvement of the LC in the mobilization of control
for action triggering. Indeed, as mentioned earlier (see part I.B.1.b), LC axons are
unmyelinated with a slow conduction velocity, around 0.9m/s (Aston-Jones et al., 1985).
Consequently, LC inputs would reach areas involved in control exertion within a little below
150ms, i.e. at the very end of the RT. Still, LC activity before action onset could influence the
action itself, for example by improving the gain of the motor system in the case of physical
effort (see part I.C.2.a) for gain effects). Alpha motor fibers, which conduct the action
potentials responsible for skeletal muscle contraction, are myelinated fibers with high
conduction velocity, around 100m/s in primates (Chipman et al., 2012), meaning that the
transmission of the motor message from the motor cortex to the muscle takes less than 20ms.
Pre-action LC activity thus appears to have a timing potentially compatible with influencing
the execution of motor actions resulting from the action triggering process via an influence on
motor cortices. Such reasoning can putatively be applied to actions involving mental effort,
such as triggering the process of making a mental calculation. Additionally, LC neurons may
be reporting to other areas on how much control has been invested in the process of triggering
the action, and consequently be part of the monitoring unit of control, or at least provide
some information to it (see part I.A.2.c). The evidence for this hypothesis is clearly insufficient
at this stage, since the task parameters encoded by LC neurons prior to action onset are only
correlated with the subjective cost of triggering the action.
Altogether, a large body of literature provides some evidence pointing in the direction
of an involvement of the LC in the action triggering process, because LC neurons activate at its
end in a magnitude reflecting the subjective cost of the action being triggered. However, the
87
timing of this activation cannot account for a role of the LC in mobilizing cognitive control for
triggering the costly actions, as it appears to be too late, predicting the end of this process just
before it actually happens and the action begins. In the next part, I review the literature
regarding a potential involvement of LC activity in facilitating action monitoring, and argue
that this may be the function of such pre-action activation.
6. LC activity and action monitoring
Following the triggering of an action, subjects have to monitor its consequences both
in the short and long term (e.g. error or success, reward, punishment, fatigue). In that sense,
a crucial component of action monitoring is error monitoring. Post error slowing is an increase
of RTs following errors, reflecting behavioral adjustments following error detection (Rabbitt,
1979). Murphy and colleagues (2016), showed that the size of the error-evoked pupil response
could predict the magnitude of the post-error slowing. Considering the positive relationship
between LC activity and pupil dilation (see part I.B.2.b), this relationship between pupil
responses and post-error slowing could reflect a role of LC activity in error monitoring and
adjustments in response to the error, such as increased cognitive control exertion and
therefore response slowing (see part I.A.2.f)(1) for the relationship between RT and cognitive
control). Sellaro and colleagues (2015) showed that transcutaneous vagus nerve stimulation
(tVNS), a technique thought to activate the LC (Ariely et al., 2009; Dorr and Debonnel, 2006;
Frangos et al., 2015) enhanced post-error slowing.
Action monitoring and in particular error monitoring have been linked to the activity
of the ACC (e.g. Botvinick et al., 2004; Dehaene et al., 1994; Gehring and Knight, 2000; Luu and
Tucker, 2001; van Veen et al., 2001; van Veen and Carter, 2002). The ACC receives numerous
LC axons (e.g. Schwarz et al., 2015). Selective stimulation of LC to ACC projections increase
glutamatergic synaptic transmission and neuronal excitability in the ACC (Koga et al., 2020).
Moreover, the correlation of activity of single ACC neurons (spike count) may have increased
in case of high signal to noise ratio in LC activity (low baseline and strong responses to events),
and decreased in presence of low signal to noise ratio (Joshi and Gold, 2020, preprint). These
findings tend to indicate that task-related LC activity could be increasing the efficacy of
synaptic transmission in the ACC.
To summarize, LC activity increases, especially when this activity has a high signal-to-
noise ratio, have been associated to facilitated action execution and especially error
88
monitoring. In that frame, the burst of activity exhibited by LC neurons prior to action onset
could serve to enhance action monitoring and facilitate error detection. Additionally, the
pattern of activation of LC neurons in response to fixation breaks (errors by disengagement
from the task), featuring a burst of activity especially strong before re-engagement in the task,
reported by Jahn and colleagues (2020), could constitute a mechanism through which LC
neurons participate in action monitoring and subsequent adjustments.
In this part, I first detailed evidence of the involvement of LC activity in a large range
of non-goal directed costly cognitive processes: arousal, stress and attention. In the context
of goal-directed behavior, moderate LC activity has been linked to performance optimization
in tasks demanding cognitive control exertion, for working memory updating, response
inhibition and task-switching, although high LC activity leads to task disengagement and
exploration (for inverted U shape curve of the relation between LC activity and performance,
see part I.C.2.b). In the specific context of decision-making, LC activity influences action
selection, but that influence is very probably indirect, via alterations of the perception of effort
(mental or physical) when LC activity is artificially modified. Moreover, LC activity seems to
participate in the action-triggering process, and in particular report the difficulty of the action-
triggering process at its end. Additionally, moderate increases in LC activity lead to enhanced
action monitoring, and in particular error monitoring. Taken together, these roles of the LC in
decision-making could indicate a role of the LC in regulating effort allocation and production.
89
C. Mechanisms for the effects of LC activity on target
areas to regulate cognitive processes, and implications
for effort processing
“...adequate characterization of the functional properties of the noradrenergic system
requires that its impact be considered with regard to other neural systems projecting on the
same target neurons.”
Foote and Morrison, 1987
In the previous part, I reviewed evidence towards an involvement of the LC/NA system
in effort processing, and especially in monitoring the control exertion process and facilitating
action execution. In this part, I shall review potential mechanisms: molecular mechanisms via
metabolic adjustments, short-term and longer-term electrophysiological adjustments, i.e.
signal-to-noise ratio enhancement and synaptic plasticity promotion, respectively, and
network adjustments, i.e. connectivity changes across brain regions.
1. LC activity modulates brain metabolism: NA and response
to high metabolic demands
Cognitive control and physical effort are perceived as costly and aversive. The cost of
cognitive control has been hypothesized to be the consequence of the depletion of a specific
resource, which could be metabolic (see part I.A.2.e)(1)). Specifically, Christie and Schrater
(2015) have proposed the possibility that the avoidance of cognitive control exertion could be
an adaptive mechanism aiming at sparing astrocytic glycogen, a storage polymerized form of
glucose. Additionally, Matsui and colleagues (2017) have demonstrated that physical effort
exertion depletes astrocytic glycogen reserves, and that this depletion is critically linked to
fatigue following prolonged physical effort. Together, this literature resonates with the
extensive findings regarding the effect of noradrenergic release on astrocytic glycogenolysis,
i.e. release of lactate from glycogen, thus providing nearby neurons with metabolic substrate
(see Figure 15).
90
NA release in the brain signals high metabolic demand following increased brain
activity (Magistretti and Morrison, 1988). Via extrasynaptic transmission from LC axon
varicosities located close to glial cells throughout the cortex (Fuxe et al., 2012), NA release
modifies brain metabolism by actions on glycogen-containing astrocytes, from early stages of
development throughout adult life (Cohen et al., 1997; Gibbs, 2015; Gibbs et al., 2010; Hertz
et al., 2010; Müller, 2014; O’Dowd et al., 1995; Quach et al., 1978). First, regarding short-term
effects, NA induces lactate release from astrocytes by increasing glycogenolysis (Magistretti,
1988; Quach et al., 1978; Sorg and Magistretti, 1991). This effect is mainly dependent on β
adrenergic receptors although α adrenergic receptors may play some part (Sorg and
Magistretti, 1991). NA stimuli can mobilize more than half of the glycogen reserves, providing
fast and strong energy surges. On the longer time-scale i.e. several hours, NA participates in
the replenishment of glycogen reserves beyond what can be explained by the initial
glycogenolysis, by modulating the synthesis of proteins necessary for glycogen synthesis (Sorg
and Magistretti, 1992), specifically by promoting the transcription of genetic material related
to glycogen synthesis such as PTG (protein targeting glycogen, Allaman et al., 2000). Such
metabolic effects have strong consequences on cognitive process, such as memory
consolidation (Gibbs, 2015; Gibbs et al., 2010).
Overall, such accounts of the role of NA in astrocytic glycogen immediate
transformation into lactate and long-term replenishment supports the idea of an implication
of the LC/NA system in responding to higher brain metabolic demands. These accounts
regarding the role of NA in rendering metabolic substrates (lactate) available offer a potential
mechanism for the hypothesis that LC pre-action activity may participate in action execution
and action consequences monitoring.
2. From LC facilitation of sensory processing and plasticity to
gain theory
a) Facilitation of sensory processing
LC neurons have been shown to respond to unexpected sensory stimuli (Bouret and
Richmond, 2009; Bouret and Sara, 2004; Grant et al., 1988; Vankov et al., 1995, see Figure 20)
and task-relevant cues (Aston-Jones et al., 1994; Aston-Jones and Bloom, 1981; Bouret and
91
Richmond, 2015; Bouret and Sara, 2004, 2002; Jahn et al., 2020; Kalwani et al., 2014; Varazzani
et al., 2015, see Figure 35).
Cognitive control has been theorized to involve modifying the identity of the
information processed and intensity of processing (Shenhav et al., 2017). NA implication in
sensory processing has long been hypothesized based on indirect or in vitro proof (e.g.
Ciombor et al., 1999; Halliday et al., 1994; Heath et al., 2006; Turetsky and Fein, 2002), but
only in the recent years have causal manipulations confirmed this effect and allowed
identifying its mechanism. Such studies used electrical stimulation of the LC or modifications
of NA brain levels with direct application of NA in target areas, or pharmacological agents,
such as atomoxetine (increasing NA) or clonidine (decreasing NA), and simultaneously
recorded neuronal activity in areas of interest (electrophysiological recordings or fMRI).
(A) (B)
Figure 35: LC neurons respond to task-relevant cues.
(A) Raster plot of the activity of one single-unit recorded from the LC in a Pavlovian task,
aligned on the onset of task cue. The line of circles indicates when the red dot requiring ocular
fixation was displayed. Note that for each trial, response consisted in one to three spikes
followed by an inhibition period. (adapted from Bouret and Richmond, 2009)
(B) Activity of LC neurons around cue onset in a force-reward task. Top: raster plot of example
neuron responding to cue onset. Blue line is the spike density plot. Note that for each trial,
response consisted in one to three spikes followed by an inhibition period. Bottom:
92
representation of mean rate per neuron across time (sliding window procedure). Neurons
consistently responded to cue onset yet with variable latencies. One line corresponds to the
activity of one single unit (adapted from Varazzani et al., 2015).
Specifically, NA release in sensory cortices (auditory, visual, olfactory and
somatosensory), and other regions critical for treating sensory information (e.g. the dorsal or
lateral geniculate nucleus or the thalamus), enhances the signal to noise ratio (Ciombor et al.,
1999; Devilbiss et al., 2006; Foote et al., 1975; Hurley et al., 2004; Kossl and Vater, 1989; Segal
and Bloom, 1976; Waterhouse et al., 1998) and gates sensory inputs6 (Ciombor et al., 1999;
Jiang et al., 1996; Waterhouse et al., 2000). Such findings were reviewed by Hurley and
colleagues (2004). Figure 36A illustrates the effect of LC stimulation on the firing pattern of
thalamic neurons.
These effects have been attributed to the enhancement of excitatory responses via α1
receptors associated to an enhancement of inhibitory responses by β adrenergic receptors
(Waterhouse et al., 1988, 1980; Waterhouse and Woodward, 1980). Additionally, NA appears
to enhance the specificity of the coding of sound frequencies in auditory neurons, making the
auditory system more finely tuned (Manunta and Edeline, 2004). In the somatosensory cortex,
stimulating the LC shortly before touch stimuli, in addition to increasing the signal to noise
ratio, enhanced the temporal precision of the evoked response, i.e. the latencies to the
evoked responses were less noisy (Lecas, 2004), as can be seen in Figure 36B. This increase in
spike timing precision could facilitate stimulus encoding (Dan and Poo, 2006; Lestienne, 2001;
Tiesinga et al., 2008).
Overall, NA improves the sensitivity to incoming stimuli and their processing (Devilbiss
et al., 2006; Foote et al., 1975; Gelbard-Sagiv et al., 2018; McCormick, 1989; Navarra et al.,
2008; Rogawski and Aghajanian, 1980; Vazey et al., 2018). LC neurons selectively respond to
salient and task-relevant stimuli, but these responses could be too late to facilitate the
processing of these stimuli. For instance, Lecas and colleagues (2004) had to stimulate the LC
200ms prior to stimulus onset to observe significant effects on stimulus processing in the
somatosensory cortex (Figure 36B). Consequently, in normal conditions, NA would only
facilitate the encoding of stimuli presented after salient or task-relevant stimuli. Therefore,
6 Sensory gating is the process that filters irrelevant or redundant stimuli (Sofuoglu et al., 2014).
93
NA release in response to task cues and events (such as actions) could lead to improvements
in the processing of their consequences and therefore enhance task-associated information
processing. It is interesting to note that pre-action peaks of LC activity recorded in LC neurons
occur around 200ms prior to action onset (Bouret and Sara, 2004; Clayton et al., 2004;
Varazzani et al., 2015), which means that they have an ideal timing for enhancing the
processing of action-related sensory stimuli, critical for action monitoring and adjustments.
Overall, the facilitation of processing by LC activity could account, at least in part, for the
effects of NA on performance in a various range of tasks and especially on action monitoring.
(A) Devilbiss et al., 2006, Figure 3, lines 1&2 (B) Sara, 2009, Figure 4A
Figure 36 : Effects of LC stimulation on activity of simultaneously recorded thalamic neurons
and somatosensory neurons.
(A) The LC was electrically stimulated at different frequencies and activity in the thalamus
following sensory stimulation was recorded. Peri-stimulus time histograms of thalamic activity
(aligned on whiskerpad stimulation) show increased responses to the stimulus and decreased
baseline, i.e. increased signal-to-noise ratio. NB: The superimposed numbers indicate the
maximum response observed (rate at peak) (adapted from Devilbiss et al., 2006).
(B) The LC was stimulated 200ms prior to tactile paw stimulation. Activity in somatosensory
neurons was recorded, and is represented in these rasters aligned on paw stimulation. The top
raster represents the response of the somatosensory neurons in the absence of LC stimulation
and the bottom raster represents the activity of the somatosensory neurons after LC
stimulation. Evoked responses are more reliable and their latencies are less noisy after LC
stimulation (adapted from Sara, 2009).
Although such effects of facilitation of neuronal integration have (to our knowledge)
mainly been demonstrated in sensory cortices and other areas involved in sensory processing,
several studies have pointed towards a more generalized effect on several other brain areas
involved in other categories of functions. For instance, increased inputs from LC to cerebellum,
obtained by electrical stimulation of the LC, caused inhibition of baseline spontaneous activity
of Purkinje cells (Freedman et al., 1977; Hoffer et al., 1973) and promoted synaptic plasticity
94
(Carey and Regehr, 2009). In motor area M1, blocking NA reuptake with atomoxetine
increased neuronal excitability and improved motor performance (Plewnia et al., 2004).
As far as mental effort is concerned, several studies (Li et al., 1999; Wang et al., 2007)
provided some evidence for at least partial signal-to-noise ratio modulation of PFC neuronal
responses in the context of spatial working memory demands. These teams first identified PFC
neurons displaying spatial encoding of cue location7 during delay periods during which this
location had to be retained in working memory, and applied α2 agonist guanfacine or α2
antagonist yohimbine onto them. Guanfacine at moderate doses enhanced and yohimbine
suppressed responses of PFC neurons during the delay period of the task (Figure 37A-C),
suggesting that in this context NA enhances the signal in PFC neurons involved in spatial
working memory. These studies however provide no evidence for a decrease in the noise in
the signal. Note that Wang and colleagues (2007) additionally provide evidence that effects of
α2 receptor stimulation on working memory critically depend on a pathway involving
inhibition of effects of cyclic adenosine monophosphate (cAMP). cAMP opens HCN channels
(hyperpolarization-activated cyclic nucleotide gated cation channels), located on the dendritic
spines of neurons in the superficial layers of the PFC, and the opening of HCN channels shunts
nearby synaptic inputs, therefore weakening their effect. α2 receptors are located near HCN
channels and could probably modulate their opening, increasing the efficacy of the inputs
(Figure 37D).
(A) Wang et al., 2007, Figure 1A (B) Wang et al., 2007, Figure 2E left panel
(C) Wang et al., 2007, Figure 2F left panel (D) Arnsten and Pliszka, 2011, Figure 3
Figure 37: In a spatial memory task, α2 agonist guanfacine enhanced while α2 antagonist
yohimbine decreased spatial-related responses of PFC neurons during memory retain period.
(A) Schematic representation of the task used. Monkeys had to fixate a central dot on the
7 Note that PFC neurons displayed responses with a preferred direction, e.g. some neurons only responded when the cue was presented in a certain direction (angle from central fixation point). When mentioning suppression of enhancement of responses of LC neurons I therefore focus on trials in which the cue whose position had to be retained was presented in the preferred direction of the studied neurons.
95
screen. During fixation, a cue was displayed for a short amount of time. After a delay, monkeys
had to perform a saccade onto the position of the cue (adapted from Wang et al., 2007).
(B,C) Activity of PFC neurons (in the case of the cue being presented in their preferred direction)
in control condition (blue) and after (B) guanfacine or (C) yohimbine local (adapted from Wang
et al., 2007).
(D) Proposed mechanism for the effects of α2 stimulation in the PFC on synaptic efficacy. In
absence of NA, cAMP causes the opening of HCN channels (blue), which weakens the inputs to
the neurons by shunting them. In presence of NA, α2A adrenergic receptors are stimulated,
inhibiting the effects of cAMP and therefore the opening of the HCN channels. NA stimulation
therefore inhibits the shunting of the synaptic inputs and leads to increased synaptic efficacy
(adapted from Arnsten and Pliszka, 2011).
Altogether, these results indicate that the effect of improving synaptic transmission
efficacy, via increasing signal-to-noise ratio, referred to as neuronal gain modulation, could
therefore exist at a larger scale in the brain.
To summarize, LC activity and NA discharge onto many target areas rapidly causes
modifications in synaptic transmission: signal-to-noise ratio, gating, and even filtering of
information. Although these effects have mostly been demonstrated in areas responsible for
sensory processing, they are also present in various other regions, including cognitive control-
related regions such as the PFC. In the next part, I review the major theories providing global
theoretical frames for these effects.
b) Overarching theories of short-term effects of NA on
neuronal activity: brain states and gain modulation
NA brain concentrations and LC discharge have long been known to follow the Yerkes-
Dodson law, initially devised to explain the inverted U shape relationship between arousal or
stress and performance (Yerkes and Dodson, 1908). Due to different affinities of the different
adrenergic receptors present in the brain, Arnsten (2000) has hypothesized that this pattern
is in fact due to the engagement of qualitatively different receptors for quantitatively different
96
concentrations of NA, with moderate levels of NA engaging high-affinity α2 adrenergic
receptors and higher levels engaging α1 and β adrenergic receptors. These receptors have
opposite effects on executive functions with α2 receptors improving and α1 and β receptors
impairing cognitive functions. Note that this is believed to involve differential modulation of
PFC functioning (Robbins and Arnsten, 2009). Aston-Jones and Cohen (2005) attributed this
relationship to a balance between two modes of firing of the LC: tonic and phasic (Figure 38).
(A) Aston Jones and Cohen, 2005, Figure 2 (B) Usher et al., 1999, Figure 1A
Figure 38: Yerkes-Dodson relationship between LC discharge and performance.
(A) Schematic representation of the theory by Aston-Jones and Cohen. At low rates, subjects
tend to be inattentive to the task. At intermediate rates, task engagement and performance
are optimal, and at higher rates, subjects disengage from the task and become distractible
(adapted from Aston Jones and Cohen, 2005).
(B) Activity of an example single-unit of the LC (top plot, rate across time) in a vigilance task
and its relation to task performance (frequency of false alarms i.e. FA rate, bottom plot).
Increased baseline LC activity is associated with increased false alarm rate, i.e. lower
performance and increased distractibility (adapted from Usher et al., 1999).
The tonic mode is described as a high baseline period with low signal-to-noise ratio
(weak responses to salient or task-related events), associated to high distractibility and
disengagement from current tasks. The phasic mode, conversely, corresponds to periods of
low baseline, strong task- or salient stimuli-elicited responses and therefore high signal-to-
noise ratio, and this mode is thought to be associated with optimal performance. Studies
tracking pupil diameter during exploration-exploitation tradeoffs provided some arguments
in favor of this interpretation of the Yerkes-Dodson law (Gilzenrat et al., 2010; Jepma and
Nieuwenhuis, 2011, see part I.B.4). Note that although Aston-Jones and colleagues reported
observing such modes during in vivo recordings of LC single-units in awake behaving animals,
to our knowledge, no other team has succeeded in reproducing their observation, and this
account is now strongly discussed. Consequently, reference to these two modes will be
97
avoided, and replaced by reference to low versus high LC activity or low versus high
noradrenergic tone, associated, respectively, with high versus low signal-to-noise ratio.
Note that electrotonic coupling of LC neurons (Christi and Jelinek, 1993; Christie et al.,
1989; Ishimatsu and Williams, 1996), via electrical synapses (gap junctions), could account for
such differences in signal to noise ratio (Aston-Jones et al., 1999; Usher et al., 1999). Coupling
would average the uncorrelated neuronal noise across the population, leading to increased
signal-to-noise ratio in the LC (decreased baseline firing and increased task-related firing).
However, the existence of the necessary gap junctions between LC neurons in adult humans
and other animals is widely discussed. Indeed, electrotonic coupling is thought to be
dependent on the existence of gap junctions between astrocytes and LC neurons, but accounts
of such existence of gap junctions were made using brains from juvenile individuals (Alvarez-
Maubecin et al., 2000). In adult brains, studies of such coupling provided evidence for weak
coupling at best (Ishimatsu and Williams, 1996; Travagli et al., 1995). Alvarez and colleagues
(2002) have shown ex vivo that, in the adult rat brain, contrary to that of younger specimens,
firing synchrony and coupling between LC units is conditioned by the firing rate, with coupling
only for low rates of firing. This coupling is thought to be dependent on α2 adrenergic
receptors (Christie, 1997; Ennis and Aston-Jones, 1986), with a high NA affinity that could
account for the necessity for low LC rates for coupling in the adult brain. Considering that high
performance is generally associated to low baseline rates of LC neurons, with higher rates
causing disengagement from task and low performance, frequency-dependent coupling of LC
neurons could be one of the mechanisms involved in performance optimization.
Nonetheless, low LC discharge rates are associates to drowsiness, while intermediate
rates (about 2Hz) are usually associated to optimal performance, and high rates (about 3Hz)
with distractibility and poor performance at tasks (Rajkowski et al., 1994). Such observations
have led Berridge and Waterhouse (2003) to formulate the hypothesis according to which LC
discharge, depending on its intensity, modulates cognition by putting the brain into a certain
state, e.g. the waking state. Within this state, the LC is then hypothesized to modulate the
salient information processing, via a combination of modifications of synaptic excitability,
susceptibility to plasticity, etc.
Aston-Jones and Cohen (2005) extended this idea, formulating the adaptive gain
theory, which is, still today, one of the most influential overarching theories for the global
98
function of the LC-NA system. Servan-Schreiber, Printz and Cohen (Servan-Schreiber et al.,
1990) initially developed the theory of cognitive function improvement via neuronal gain
increase, although they did not offer a clear hypothesis for the mechanism underlying such
gain increases8. This team, using the example of the signal-detection system, modelled the
response of individual single units to sensory inputs as a logistic function, called activation
function, relating the strength of the input to the probability of neuronal firing in response to
the input (Figure 39).
Servan-Schreiber et al., 1990, Figure 1B
Figure 39: The activation function, relating the input of a neuronal system (x-axis, Net input)
to the probability of firing in response to the input (y-axis, activation).
The function belongs to the family of logistic functions (sigmoidal shape). If gain (G) increases,
the temperature of the sigmoid curve increases. (adapted from Servan-Schreiber et al., 1990).
At every time point, theoretical single units receive inputs from other neurons,
excitatory (positive) in presence of the target and inhibitory (negative) or null in the absence
of the target. The unit’s net input is the current input minus the mean noise, and the unit fires
to signal the presence of the target only if a certain threshold is reached. Consequently,
increased noise or weaker inputs lead to lesser probabilities of the unit correctly categorizing
the target as present or absent. At the level of a group of single units, the result on behavior
would be target misses (wrong absence categorization) or false alarms (wrong presence
categorization). The temperature of the sigmoid of the activation function of individual units
can be modified by adjusting gain (Figure 39), which, at the level of a chain of units,
consequentially increases detection performance of the target.
Based on the findings regarding synaptic effects of LC activity on target areas, Aston
Jones and Cohen (2005) applied the model from Servan-Schreiber, Printz and Cohen (Servan-
Schreiber et al., 1990) to LC activity and NA effects on the brain. They summarized all effects
8 They only showed that their model fitted satisfyingly methylphenidate (global central nervous system stimulant) induced performance increases in the continuous performance task (series of letters flashed sequentially, subjects have to report when two consecutive letters were identical).
99
of LC discharge under the notion that NA causes increases in gain in target structures. Below
the optimal gain point (maximum of the Yerkes-Dodson inverted U shaped curve), increases
in gain are beneficial for task performance, i.e. increases in LC firing beneficially increase gain
in target areas, resulting in better task performance. In that sense, moderate LC activity
promotes exploitation. Beyond that point (in the downward part of the inverted U curve),
increases in LC activity are no longer beneficial for task performance, as they increase gain so
much that they cause distractibility because the system becomes sensitive to task irrelevant
stimuli. In that sense, high LC activity promotes exploration. Note that such inverted U shape
of LC activity on the signal-to-noise ratio, i.e. gain, of the activity of target areas can be
visualized in the pattern of response of the top thalamic neuron of Figure 36 to increased LC
tone.
Reconciling this account with our hypothesis of an implication of the LC/NA system in
effort processing requires considering it in the context of tasks requiring cognitive control
exertion, which, in fact, only constitutes a particular case of the larger proposed rule. For
instance, in the case of decision-making, optimal neuronal gain could allow better processing
of relevant information while efficiently ignoring irrelevant information. In addition, this
theory offers a theoretical frame and potential mechanism for regrouping accounts of the
implication of the LC/NA system in effort processing and costly cognitive processes in general
with findings regarding an implication in behavioral flexibility.
In the particular case of decision-making, and specifically of two-alternative forced
choices, Shea-Brown and colleagues (2008), generalizing models theorized for target
identification (Gilzenrat et al., 2002; Usher et al., 1999; Usher and Davelaar, 2002), devised a
multilayered neuronal model of decision-making. In brief, they created a two layers model in
which the first layer, a sensory layer, encoded task information, and transferred that
information to a second layer of response accumulators. When sufficient evidence was
accumulated in favor of one option in the accumulator layer, the response was triggered.
Additionally, when the accumulated evidence reached a certain threshold, which was below
the response threshold, gain was adjusted in both layers, i.e. gain was adjusted when nearly
enough evidence had been accumulated, leading to an acceleration of the evidence
accumulation towards the response. They then compared the maximum reward rate
obtainable in the absence of gain modulation compared to reward rate in a model with gain
modulation. If model parameters, including gain, were optimized, reward rate increased by
100
14 to 24% compared to a two-layers model with no gain adjustments. Note that, by
construction, in this model, gain was increased at the very end of the RT, and speeded up the
response, and such a pattern could be compatible with the timing of pre-action LC activations
(see part I.B.5.b).
c) Extending gain theory in time: LC activity and
plasticity promotion
At larger temporal scales than adjustments of signal-to noise ratio of individual
neurons, LC activity has been shown to influence synaptic plasticity. Plasticity is the ‘activity-
dependent modification of strength or efficacy of synaptic transmission of preexisting
synapses’ (Citri and Malenka, 2008), via long-term potentiation and/or depression (LTP and
LTD). Plasticity allows the brain to adjusts its activity and connectivity, and therefore
subsequent behavior, and is particularly critical for memory formation.
In addition to signal-to-noise ratio increasing effects, plasticity-promoting effects of NA
were observed. As indicated by their names, LTD and LTP are long-lasted, and NA effects on
neuronal gain could be prolonged in time by these phenomena. Studies reporting NA-
dependent plasticity were initially performed in sensory networks, visual (Pettigrew, 1982;
Pettigrew and Kasamatsu, 1978) and olfactory (Sullivan et al., 1989) (for exhaustive recent
review, see Perez, 2020). NA was shown to promote plasticity by inducing LTP, in the CA3
region of the hippocampus (Hopkins and Johnston, 1984; Stanton and Sarvey, 1985), as well
as the CA1 and the dentate gyrus (Harley and Sara, 1992; Heginbotham and Dunwiddie, 1991;
Neuman and Harley, 1983; Segal and Bloom, 1974; Stanton and Sarvey, 1985). More recently,
these findings were complemented by studies showing that NA can induce LTD in the
hippocampus via fixation on β adrenergic receptors (Hagena et al., 2016; Lemon et al., 2009;
Straube et al., 2003), and in the visual cortex via fixation on α1 and β adrenergic receptors
(Salgado et al., 2012). The balance between NA favoring LTD or LTP appears to be
concentration-dependent and spike-timing dependent (Lemon et al., 2009; Salgado et al.,
2012). Note that the co-release of DA with NA (see part I.B.1.c) is thought to have a critical
role in plasticity promotion, at least in the hippocampus (Sonneborn and Greene, 2021).
101
In addition, various studies have demonstrated that increasing brain NA levels
naturally (stress exposure), pharmacologically (α2 antagonist or β agonist drugs) or directly
(infusions of NA in target areas) produce robust and fast increases in transcription of genes
and protein synthesis critical for long term plasticity (e.g. Bing et al., 1992, 1991; Gubits et al.,
1989; Maity et al., 2020; Stone et al., 1997, 1991; Walling and Harley, 2004).
Regarding the mechanisms regulating gain and plasticity effects of LC activity, the
GANE model (glutamate amplification of NA) posits that NA concentrations in LC target areas
are modulated locally by glutamate, with high glutamate in targets causing enhancement of
NA release by stimulation of LC synaptic bouton NMDA (N-methyl-D-aspartate) receptors,
thus creating ‘NA hotspots’ (Mather et al., 2016). In the hotspots, NA would facilitate the task-
relevant processes and representations. Around these hotspots, lateral inhibition would be
enhanced, such that irrelevant processes and representations would be inhibited.
Additionally, in the absence of NA hotspots, NA levels would be sufficient only to activate
inhibitory receptors with high affinity. In that sense, the GANE theory can be understood as a
larger scale generalization of the gain theory. While the adaptive gain theory focuses on the
signal-to-noise ratio of individual units, the GANE theory proposes that NA neurotransmission
is also associated to a change in the “signal-to-noise ratio” of larger neuronal networks and
processes, enhancing relevant and inhibiting irrelevant computations and neuronal
connections.
Altogether, these accounts suggest that not only is the effect of NA on synaptic
transmission efficacy probably generalized in space, i.e. throughout the brain, but also in time,
as NA induces long-term circuitry adjustments by promoting synaptic plasticity. In the next
part, I focus on a larger spatial scale, and review evidence for connectivity adjustments and
network rearrangements following NA release.
d) LC activity and brain connectivity and networks
Bouret and Sara (Bouret and Sara, 2005; Sara and Bouret, 2012) proposed an
overarching theory for the role of the LC/NA system on behavior: network reset for behavioral
shifts (Figure 40), later extended by Corbetta and colleagues (2008). In that context,
responding to a cue or making a choice after a deliberation period constitutes a behavioral
102
shift, as well as switching from exploitation to exploration. Thus, the finding that the LC/NA
system is critically involved in set-shifting, and that LC neurons respond to stimuli requiring
behavioral adjustments (cues) and prior to behavioral changes (responses) advocates for such
a role in behavioral shifts. Recent brain connectivity studies have provided evidence for such
brain network reset in response to LC activation (reviewed Corbetta et al., 2008).
Figure 40: Network reset theory (adapted from Bouret and Sara, 2005).
In a given state of a task (e.g. set of task rules), the network serving to solve the task is in a
certain configuration, with certain units (circles) activated (green) and certain silent (grey).
Following a shifting stimulus (e.g. cue or change in rules), LC neurons activate and NA is
released. NA acts on the network and modifies the pattern of response of its units. Afterwards,
a different subpopulation of the network is active.
Complementarily, Yu and Dayan (2005) proposed that NA could serve to signal
unexpected uncertainty, i.e. brutal changes in the environment, for instance task
contingencies, and therefore trigger adjustments to such changes. They opposed this role to
a putative role of acetylcholine in signaling expected uncertainty (e.g. known reward
probabilities in a probabilistic task or standard deviation in the distribution of outcomes). Note
however that following experimental studies only managed to demonstrate a role of NA in
uncertainty signaling in general (Jepma et al., 2018; Jepma and Nieuwenhuis, 2011; Nassar et
al., 2012; Payzan-LeNestour et al., 2013). Nassar and colleagues (2012) specifically linked
expected uncertainty (uncertainty in the expected mean of the distribution of the outcome)
and unexpected uncertainty (probability of a sudden change in the distribution of outcomes)
103
to pupil diameter around outcome onset, and specifically with pupil diameter average and
pupil diameter change, respectively, during this epoch.
Coull and colleagues (1999) provided early accounts of the involvement of LC activity
in network reconfiguration throughout the brain. They showed that, during an attentional
task, α2 agonist (clonidine) administration decreased the functional strength of the
connections between areas critical for the task (from and to the visual cortex and from cortex
to thalamus) while connectivity with the parietal cortex increased. Additionally, stress, which
is associated to LC activation, has long been known to cause extensive network adaptations
throughout the brain, with increased activation in the salience network (Hermans et al., 2011;
van Marle et al., 2010), promoting vigilance and altering cognitive control capacity (Corbetta
et al., 2008; Hermans et al., 2014). Note that this effect of stress on functional connectivity is
blocked by β receptor antagonist (propranolol) administration, and therefore probably
dependent on β receptors.
Recent accounts of the relationship between the LC/NA system and network
reorganization have involved assessing functional connectivity following modifications of the
activity of the LC or the availability of NA. Functional connectivity refers to the temporal
correlation between the BOLD MRI signal voxels or brain areas (Fox and Raichle, 2007). Guedj
and colleagues (2017b) report that atomoxetine administration caused resting state
functional connectivity reorganization, with strong changes relating to the connectivity of the
fronto-parietal attention network to other brain areas. These findings indicate that LC
activation effects on functional connectivity could be critical for attention. This study however
focused on resting state connectivity, i.e. outside of task demands, a gap filled by several other
teams.
Atomoxetine administration was found to cause decreased functional connectivity in
the default-mode network, a functional connectivity pattern observed outside of tasks, with
connectivity negatively correlated to task demands (Brink et al., 2016), suggesting that NA
could critically modify brain connectivity to adjust to task demands. In Parkinson’s disease,
atomoxetine administration was shown to increase resting-state functional connectivity
between the dorso-lateral PFC and the ACC, and the strength of this increase correlated with
the degree of cognitive improvement observed in the patients (Borchert et al., 2016). In the
context of cognitive control exertion, Hernaus and colleagues (2017) demonstrated that,
104
during a working-memory task, atomoxetine administration increased the functional
connectivity between various brain areas. The increase of connectivity between the insula and
the dorsolateral prefrontal cortex was of particular interest as, first, it only appeared in case
of high, but not moderate working memory load, and, second, the magnitude of the increase
was predicted by individual working memory capacity.
Pharmacological modification of NA availability and LC activity are not entirely specific.
Zerbi and colleagues (2019), conversely, very specifically modified LC activity using DREADDs.
They confirmed the previous findings regarding changes in functional connectivity in several
brain connectomes following stress exposure or enhanced NA neurotransmission. These
effects were driven by α1 and β1 and, to a lesser extent, α2 adrenergic receptors. Note that
such a pattern of receptor involvement may have been due to the strong activation of LC
neurons caused by the chemogenetic manipulation, mimicking stress exposure, considering
the much higher affinity of α2 adrenergic receptors compared to α1 and β1 (Arnsten, 2000).
Similar receptors may be involved in putative network reconfiguration to switch from
exploitation to exploration. Conversely, different receptors (possibly mostly α2 adrenergic
receptors) may be involved in the putative connectivity modifications happening in the case
of inhibitory control exertion and in those observed in the context of enhancements of NA
neurotransmission. Indeed, inhibitory control and working memory improvements are
observed for small pharmacological increases in NA brain concentrations, while set-shifting
requires stronger NA increases, which are detrimental for working memory and inhibitory
control (Newman et al., 2008).
These effects of LC activity on large-scale neuronal networks could be the brain-wide
consequence of gain modulation, possibly through coherence of the gamma rhythm9 (Guedj
et al., 2017a) combined to LC modular organization. Depending on the context, different
modules of the LC are activated, and may assemble into functional arrays relaying relevant
information. The output of these neurons would adapt the neuronal gain in different target
structures, to promote adaptive behavior, resulting in coherent gamma rhythm across them,
constituting a synchronous network capable of better information transmission (Fries, 2005;
Nir et al., 2008; Womelsdorf et al., 2007). In that frame, effort, mental or physical, is merely
an example situation in which LC activity could modulate the neuronal and network properties
9 Gamma rhythm is a pattern of neuronal oscillation with a frequency comprised between 30 and 80Hz (Whittington et al., 2011).
105
in order to facilitate task completion. Cognitive control exertion depends on a specific
network, distinct from the salience network (Seeley et al., 2007). One can hypothesize that
specific environmental situations and behavioral states associated to cognitive control
exertion would lead to specific LC ensembles releasing NA in target areas involved in this
network, probably in amounts sensibly lower than those that cause salience network
enhancement. The same is probably true for physical effort.
D. Summary and problematic
First of all, effort, both mental (cognitive control) and physical, is aversive, i.e.
constitutes a decision variable that drives decision-makers away from high-cost options.
Additionally, mobilizing effort is beneficial for task performance, i.e. for goal attainment. For
instance, in the context of decision-making, cognitive control facilitates information
processing and inhibition of response biases or automatic responses. Effort allocation follows
a tradeoff between effort costs and benefits, and decision-makers exert effort in a way that
maximizes the expected value of that effort exertion. Physical and mental effort may rely, at
least partly, on similar depletable resources, as cognitive control exertion decreases the
capacity for physical effort exertion and vice versa. However, the nature of the resources
involved is debated, as the effort allocation system could aim at sparing metabolic substrate
(glucose, glycogen), and/or at minimizing interactions between computations conducted in
parallel, and/or merely at avoiding opportunity costs.
The LC is a heterogeneous noradrenergic brainstem nucleus receiving inputs from and
sending axons to virtually the entire brain. NA increases signal to noise ratio in LC target areas,
increases lactate availability for neuronal metabolism and adjusts brain connectivity to task
demands. The LC/NA system is causally involved in several non-goal directed processes that
lead to increased mental and physical expenses: arousal, stress and attention. In the context
of goal-directed processes, and in particular physical effort, pharmacological decreases of LC
activity have demonstrated a causal relation between LC activity and subjective difficulty
(effort costs), a theoretical computational variable affecting choices and action execution in a
coherent fashion. LC neurons activate and the pupil dilates around the onset and execution of
costly actions. In force tasks, LC activity and pupil diameter correlate positively with the
106
required and exerted force around action execution, although the exact dynamics of encoding
of these parameters and the relative contributions of objective difficulty and exerted force to
LC activity had not been examined in details (up to now). In case of cognitive control exertion,
indirect measures of LC activity (EEG P3 component and pupil diameter) have well-described
positive correlations with mental effort exertion, while pharmacological or optogenetic
modifications of LC activity influence mental effort performance (inhibitory control, task-
switching and working memory). LC neurons respond to the presentation of task-relevant
information, and pupil diameter increases during decisions, negatively predicting the strength
of biases, i.e. the cognitive control allocated to the decision. Finally, LC neurons respond to
errors (at least commitment errors), predicting the probability to reengage, and could
therefore signal inadequate effort exertion and facilitate the required behavioral adjustments.
In coherence with that, non-specific enhancements of LC activity (tVNS) have beneficial effects
on behavioral adjustments to errors (post-error slowing).
Altogether, these results suggest a critical role of the LC/NA system in physical effort
and control processing. Specifically, we hypothesized that the LC may encode effort around
the time of its production, and signal the consequences of the effort, for instance errors, which
are consequences of insufficient effort exertion. Second, based on results of long-term
manipulations of LC activity, we hypothesized that the LC could participate in adjustments of
effort allocation strategies based the costs and consequences of past effort production. In
other words, we hypothesized, first, that the LC has a facilitative role for effort production,
and second, that it is involved in monitoring effort costs and consequences in order to adjust
later effort allocation.
First, we would therefore expect LC neurons to encode effort regardless of the nature
of the challenge at hand, physical or cognitive, i.e. to exhibit a domain-general encoding of
effort. Second, considering the slow conduction velocity of LC axons, we would expect this
online encoding of effort to have delayed consequences on later effort production rather than
on the ongoing processes. Importantly, we would not expect these delayed effort facilitation
effects to be domain-specific, i.e. we would expect LC activity to facilitate later effort
production regardless of the type of effort. Third, we would expect this signaling of effort
production to participate in the adjustment of future effort allocation, i.e. to serve as a
feedback on the cost of pursued actions, for the effort-allocation system to adjust its
decisions.
107
The activity of LC neurons of five rhesus macaques was examined in three tasks
designed to mobilize physical effort and/or cognitive control (and a control task): two one-
option tasks in Study 1, and a two-alternative choice task in Study 2. Classically, physical effort
was operationalized by differential force exertion on a grip. We expected LC neurons to be
more active at the time of action execution when a higher force was executed by the monkeys.
On the other hand, cognitive control was operationalized in two ways. First, cognitive control
is critical for overriding default courses of action (Epstein, 1994; Shenhav et al., 2013; Shiffrin
and Schneider, 1977; Sloman, 1996). We posited that monkeys would have an automatic
tendency to refuse low value trials, and that they would therefore have to exert more
cognitive control to engage in trials consisting in low value offers, and trigger the associated
actions. Under the hypothesis that LC neurons were involved in encoding the cognitive control
exerted and possibly in facilitating its production, we expected their activity at the time of
triggering actions to scale negatively with the tendency of the monkeys to refuse the offer(s).
Second, in the context of multiple alternatives, cognitive control is required to decide
accurately (Epstein, 1994; Shenhav et al., 2013; Sloman, 1996). Under the postulate that LC
activity encodes control exertion, in the two-alternative choice task, we expected LC neurons
to be more active around accurate choices.
Moreover, the delayed effects of LC activity on effort production were also explored in
a two-alternative choice task, as it had several steps (two presses separated by a binary
choice). We expected LC activity at early steps of the task to facilitate effort production (both
cognitive control and physical effort) at later stages of the trials. Finally, in that task, we also
expected LC activity to serve as a feedback signal on effort for structures deciding on future
effort allocation. Practically, we expected LC activity at the time of the action of one trial to
scale with the ongoing effort production, and to predict adjustments in effort sensitivity of
choices in the next trial. In other words, we expected that increased LC activity around the
action of one trial would lead to increased effort sensitivity of choices in the next trial.
108
II. Experimental results
A. Study 1: LC neurons encode difficulty to trigger
actions in one-option forced choice tasks
1. Locus coeruleus neurons encode the subjective difficulty of
triggering and executing actions: evidence for a role in effort
processing (Bornert and Bouret, in prep)
Abstract
The brainstem noradrenergic nucleus locus coeruleus (LC) is thought to be involved in
various costly cognitive processes: arousal, stress and attention. Recent evidence has pointed
towards an implication in effort processing and especially effort production. To assess the
dynamic relation between LC activity, effort and difficulty, we recorded the activity of 193 LC
single-units in five monkeys performing two effort-discounting tasks: a delay-discounting task,
a force-discounting task, and, as a control for the relation of LC activity with low-level sensory-
motor processes, a target-detection task. First, in the effort-discounting tasks, LC neurons
displayed a phasic activation before action onset, modulated by task parameters in directions
opposite to those observed for willingness to work, suggesting an encoding of the subjective
difficulty of triggering the actions, i.e. of the required effort production, scaling negatively with
the automatic tendency to disengage. Additionally, LC neurons encoded reaction times (RT)
over and above task parameters. Importantly, this effect was absent in the target-detection
task in which RT variations across conditions solely reflected sensory-motor differences. RTs,
and therefore pre-action LC activity, could probably be interpreted in terms of subjective
difficulty of triggering the actions and/or in terms of cognitive control produced for that
process in the effort-discounting tasks. Second, in the force-discounting and target-detection
tasks in which the action consisted in producing force on a grip, LC neurons activated during
the execution of the press. The force exerted on the grip was only encoded in the force-
discounting task, in which it did represent an effort, as it affected the probability to engage in
trials. Our data points towards an domain-general online encoding of effort requirements
(subjective difficulty) and possibly effort production by LC neurons. We argue that the timing
109
of these modulations is more likely to influence the consequence of the effortful processes
(action execution for pre-action activity, and monitoring for peri-action activity) than the
processes themselves. Altogether, our work demonstrates the involvement of LC neurons in
effort processing and suggests a mechanism for the effects of NA on effort production and
sensitivity.
INTRODUCTION
The locus coeruleus noradrenergic (LC/NA) system has been shown to be involved in
several cognitive functions: arousal (Aston-Jones, 2005; Carter et al., 2010; Foote et al., 1983),
stress (Buffalari and Grace, 2007; Cassens et al., 1981; Curtis et al., 2012; Koob, 1999; Sved et
al., 2002; Valentino et al., 1991; Van Bockstaele et al., 2001), attention (Bari et al., 2020; Clark
et al., 1987; Coull et al., 2001; Reynaud et al., 2019; Witte and Marrocco, 1997), and decision-
making (Doya, 2002; Nieuwenhuis et al., 2005; Shea-Brown et al., 2008). We recently
proposed a role in effort processing (Borderies et al., 2020; Jahn et al., 2018; Varazzani et al.,
2015).
Effort encompasses multiple notions, but we defined it as a computational variable
representing the theoretical quantity of energy required to face a challenge (Pessiglione et al.,
2018; Shenhav et al., 2017). In that sense, effort constitutes a cost, discounting the value of
rewards (Cousins and Salamone, 1994; Hull, 1943; Phillips et al., 2007; Stevens et al., 2005;
Tsunematsu, 2001; Walton et al., 2006, 2003), but increased effort production is beneficial for
performance. In the particular case of cognitive effort, individuals are thought to discount the
benefits with the costs of the intensification mental activity needed to meet task goals and
inhibit automatic responses (Kool et al., 2010; Kool and Botvinick, 2018, 2014; Shenhav et al.,
2017, 2013; Zipf, 1949). In other words, subjects allocate cognitive control rationally in order
to maximize short- and long-term utility.
In that frame, it remains unclear whether the LC/NA system contributes to effort
processing by participating in decisions about effort allocation based on the estimation of
expected effort costs, referred to here as subjective difficulty, and/or facilitates the execution
of effortful actions. The recent result that pharmacological decreases of LC activity cause
increases in effort-discounting and decreases in physical effort production (despite conserved
motor capacities, Borderies et al., 2020) is compatible with a role of NA in facilitating effort
110
production (mobilization of theoretical resources for action execution), rather than a role in
signaling effort as a decision variable. Indeed, impairing the integration of the effort dimension
of offers would probably induce a decrease rather than an increase in effort sensitivity, as was
observed for reward-based decisions in case of dopamine depletion (Le Bouc et al., 2016;
Martínez-Horta et al., 2014). In coherence with that idea, LC neurons were never found to
encode the cost dimension of offers at the time of decisions, whether the cost is a physical
effort requirement or a delay to the reward represented by a number of presses to execute
(Bouret et al., 2012; Varazzani et al., 2015). Conversely, long term impairments of effort
production would result in increased effort avoidance, comparably to what is observed when
motor capacities are altered after curare administration (Gallagher et al., 2001). Additionally,
we provided evidence that LC neurons encode the exerted force in a large epoch around
action execution in force-discounting tasks (Varazzani et al., 2015), supporting a role in online
effort mobilization for ongoing costly processes.
In the present study, we examined the relation between LC activity and physical effort
and cognitive control production, using single-unit recordings in rhesus macaques. We
hypothesized that the activation of LC neurons would signal the effort produced around the
time of its production, both in the case of cognitive (e.g. inhibiting automatic behavior) or
physical effort (e.g. executing a difficult action). By contrast, the activation of LC neurons
would not be modulated by low-level sensory-motor processes as long as these processes
were not associated with effort. In other words, we expected LC activity to dynamically predict
changes in cognitive and physical effort production (energy expenditure), computed in a sub-
second scale, thus possibly participating in effort production.
We compared LC activity across two one-option tasks manipulating cognitive control
constraints for action-triggering (force-discounting and delay-discounting tasks), and/or
physical effort constraints for action execution (force-discounting task). First, cognitive control
is critical for overriding default responses (Epstein, 1994; Shenhav et al., 2013; Shiffrin and
Schneider, 1977; Sloman, 1996). Our force- and delay-discounting tasks manipulated trial
value such that monkeys exhibited differential engagement rates across conditions. Thus, the
cognitive control allocated when triggering the actions scaled negatively with the tendency to
disengage from the conditions offered. We expected LC activity to reflect these differences in
cognitive control production around the time of action triggering. Second, in the force-
discounting task, differential engagement rates also indicated that physical effort was
111
required to execute the actions. We expected LC neurons to represent the physical effort
produced around the time of action execution. As a control for the relation of LC activity with
mere sensory-motor requirements, we used a target-detection task. Engagement rates did
not vary across conditions, indicating that the cognitive control to trigger the actions and the
physical effort to execute them did not vary across conditions. Therefore, we did not expect
LC neurons to modulate their activity across conditions at the time of action triggering and
execution. Overall, our results were compatible with our predictions.
MATERIALS AND METHODS
Animals
A total of 5 male rhesus macaques (Macaca mulatta) were included in the study; 2 in
the Delay-discounting study (Monkey L, 9kg and Monkey T, 9.5kg), 1 in the Target-detection
task study (Monkey J, 12kg), and 2 in the Force-discounting study (Monkey D, 11kg and
Monkey A, 10kg),. During testing days, they received water as reward, and on non-testing
days, they received amounts of water matching their physiological needs. The experimental
procedures for the Force-discounting task and Target-detection task studies were designed in
association with the veterinarians of the ICM (Brain and Spine institute), approved by the
Regional Ethical Committee for Animal Experiment (CREEA IDF n°3) and performed in
compliance with the European Community Council Directives (86/609/EEC). The experimental
procedures of the Delay-discounting study followed the ILAR Guide for the Care and Use of
Laboratory Animals and were approved by the NIMH Animal Care and Use Committee.
Behavior
During sessions, the monkeys squatted in a primate chair, in front of a computer screen
on which the visual stimuli of the task were displayed. For the Force-discounting task and the
Target-detection task, force grips (M2E Unimecanique, Paris, France, pneumatic for force-
discounting task, electronic for Target-detection task) were mounted on the chair at the level
of the monkey’s hands, 1 for the Force discounting task and 3 for the Target-detection task
(left, right and middle grips). For the Delay-discounting task, a touch sensitive bar was installed
on the chair at the level of the monkey’s hands. The monkeys received water reward from a
tube placed between their lips but away from their teeth.
112
Tasks
Behavioral paradigms were controlled using the REX system (NIH, MD, USA) and
Presentation software (Neurobehavioral systems, Inc, CA, USA) for the Force-discounting and
Delay-discounting tasks, and using the EventIDE software (OkazoLab Ltd, London, UK) for the
Target-detection task.
Delay-discounting task
To examine the link between LC activity and cognitive effort, we designed a task in
which sensory and motor requirements would be virtually equivalent across all options, to
facilitate the comparison of cognitive processes across conditions. Therefore, the task
involved a delay to the reward as a cost. Practically, the task involved releasing a touch
sensitive bar in a given interval of time in order to get a reward of a certain size, after a certain
delay (Fig. 1A). The reward size and delay factors were orthogonalized. There were nine
experimental conditions, corresponding to the combinations of three levels of reward size (1,2
or 4 drops of water) and three levels of delay to reward delivery (400-600ms, 3000-4200ms,
6000-8400ms, Fig. 1B). Within sessions, these nine conditions were randomly distributed with
equal probability of appearance. Each combination of reward size/temporal delay was
associated to its own visual cue, displayed at the beginning of the trial to indicate the Reward-
Delay contingency of the trial. Each trial began when the monkey touched the bar, and the
cue was then displayed for 400ms (Fig. 1A). Then, a red dot appeared, for a random duration
between 1000 and 2000ms, after which it turned green, indicating that the monkey had to
release the bar. If the monkey responded between 200 and 1000ms, the green point turned
blue, as a feedback to indicate correct performance. On correct trials, monkeys had to wait
for the delay indicated by the cue before getting the announced liquid reward. A new trial
then began after a 1000ms inter-trial interval. If the monkey failed to release the bar during
the appearance of the green point (omission), or released the bar too early (anticipation), the
trial was interrupted, all visual stimuli disappeared from the screen, and the same trial started
again.
Force-discounting task
The Force-discounting task was described in detail in Varazzani et al. (2015). In
summary, the task consisted in squeezing a pneumatic grip above a force threshold in order
to get a reward, delivered after each successful force exertion (Fig. 1B). At the beginning of
each trial, a red dot was presented on screen before the apparition of a cue. Subjects were
113
instructed to fixate their gaze on the dot throughout the trial. The cue indicated the amount
of reward at stake in the trial (3 reward levels, 1,2 or 4 drops of water) and the minimum
amount of force to exert in order to get that reward (position of the force threshold, 3 force
levels, Fig. 1C). The two dimensions were orthogonalized, therefore the task had 9
equiprobable conditions, presented in pseudo-random order. After a variable amount of time
(1500±500ms from cue display), the dot turned green (Go-signal) and the monkeys had
1000ms to initiate the action i.e. to start squeezing the pneumatic grip. If the force exerted
reached the minimum threshold instructed by the cue, the dot turned blue, and remained
blue if the force exerted remained above the threshold for 500±100ms. If the force had been
maintained above threshold for the whole period, the reward announced by the cue was
delivered. An error was registered if the monkey did not engage at all (no squeezing), engaged
(started squeezing) before the go-signal, did not press with enough force, did not maintain
the force exerted above threshold for long enough (during the whole duration of the blue
dot), or broke gaze fixation on the dot. Erroneous trials were repeated until completion.
Figure 1: Summary of the Delay-discounting, Force-discounting and Target-detection tasks
114
(A,B) Delay-discounting task.
(A) At the beginning of each trial, a cue appeared, indicating the combination of reward size
and delay. After 500ms, a red point appeared, superimposed on the cue, indicating that the
monkey had to wait. After a random interval of time (1000-1500ms), the red point turned
green, lasting 1500ms. If the monkey released the bar between 200 and 1500ms after the
onset of the green point, the point turned blue, indicating correct performance. In correct trials,
the reward announced by the cue was delivered after a random delay comprised in the interval
of time announced by the cue. If the monkey did not perform any bar release, or performed it
too early or too late, the same trial started again.
(B) The task had nine conditions, a combination of three levels of Delay (400-600; 3000-4200;
6000-8400ms) and three reward sizes. Each condition was associated to a specific cue.
(C,D) Force-discounting task.
(C) At the beginning of each trial, a red dot appeared, and the monkey had to fixate his gaze
onto it. After a random period of time, a cue appeared, indicating the combination of Reward
and Force that was offered. The monkey had maintain fixation on the red dot for the whole
duration of cue presentation (1000-2000ms). The red dot then turned green (Go-signal), and
the monkey had to trigger the response within 1000ms, i.e. press the grip to reach the
instructed force threshold. When the threshold was reached, the dot turned blue and the force
still had to remain above threshold for 400-600ms. If the force was correctly maintained, the
reward predicted by the cue was then delivered.
(D) The task had nine conditions, a combination of three levels of Reward (1,2 or 4 drops of
water) and of three levels of Force. Each combination was associated to a specific cue.
(E,F) Target-detection task
(E) Trials began with the onset of a red dot in one of nine positions on the screen. After a
random delay (800-3000ms), the red dot turned green and the monkey had to press one of the
three available grips within 1000ms. The force threshold was set just above baseline, such that
any force would be sufficient. If he pressed the grip, the dot turned blue and a Reward (5 drops
of water) was delivered. In case of omission (lack of press) or anticipated press (during red
dot), the trial was restarted.
(F) The dots could appear in nine different locations on screen, defined by three potential
horizontal and three potential vertical coordinates. Horizontal coordinates indicated which
grip had to be used: if the dot was on the left side of the screen, the monkey had to press the
115
left side grip, if the dot was on the right side of the screen, the monkey had to press the right
side grip, and if the dot was in the middle column of the screen, the monkey had to press the
middle grip.
Target-detection task
The Target-detection task was designed to control for the modulation of LC activity
across conditions that differed more in terms of sensory-motor features than in terms of value
and/or effort. The tasks consisted in pressing a grip with very little force constraint to get a
fixed reward (Fig. 1E). The force threshold was set just above baseline, such that any attempt
to press was successful. Critically, the force threshold and the size of the reward were
equivalent across all conditions, which only differed by the location of the target on the screen
and the lever to be pressed. A trial began with the presentation of a red dot. The dot could
appear in nine positions (Fig. 1F). The position of the dot was maintained during blocks of trials
of random duration (5 to 50 consecutive trials) before changing location. After a very variable
amount of time (1900±1100ms), the red dot turned green (go-signal) and the monkey had to
squeeze one of the three grips within the next 1000ms. If the dot was displayed in one of the
three positions on the left side of the screen, the monkey had to press the left grip, if the dot
was displayed in one of the three positions on the right side of the screen, the monkey had to
press the right grip, and if the dot appeared in one of the 3 central positions, the monkey had
to press the middle grip (Fig. 1F). If the monkey squeezed the correct grip, the dot turned blue
for 50ms (feedback) and the reward was delivered. An error was registered if the monkey
squeezed any grip before the go-signal, or squeezed the wrong grip during go-signal
presentation, which virtually never happened during recording sessions.
Surgical procedures
Surgical procedures were the same as previously described (Bouret & Richmond, 2009;
Varazzani et al, 2015). Briefly, the approximate location of the LC was identified using 1.5T
MRI. Under general anesthesia, a sterile surgery was performed to place the head fixation post
and the recording chamber, centered stereotaxically over the body of the LC with a ~15° angle
for all monkeys.
116
Electrophysiology
Electrophysiological recordings were performed using Tungsten micro-electrodes
(UEWLEHSM3PNM, FHC, Bowdoin, ME). The electrode was placed using a stereotaxic plastic
grid (Crist Instruments) with holes 1mm apart, inserted through a tungsten guide tube (Crist
Instruments) and lowered using a hydraulic micromanipulator (Narishige). LC neurons were
identified using classical electrophysiological criteria (Grant et al., 1988): a low rate of
spontaneous activity (below 4Hz), broad waveforms (>0.6ms for the initial peak), a modulation
of firing rate across states of vigilance and a characteristic activation-pause response to brief
auditory or tactile stimuli (e.g. hand clap). Finally, we also performed clonidine tests on a series
of representative LC units at locations where LC neurons could reliably be obtained during the
course of the experiments. For this, we injected the α2-receptor agonist clonidine (20μg/kg,
IM), and all identified LC units displayed a reversible decrease in firing rate (often close to
zero, Bouret & Richmond, 2009; Grant et al., 1988). None of the units identified as non-LC cells
(e.g. neurons from the neighboring Mes5 nucleus, which respond to jaw movement) changed
their activity after clonidine injection. 92 neurons were recorded in the Force-discounting task
(n=63 in Monkey D and n=29 in monkey A), 75 in the Delay-discounting task (n=52 in Monkey
T and n=23 in Monkey L), and 26 in the Target-detection task (monkey J).
Data analysis
Analyses of the data from the Force-discounting and Target-detection tasks were
performed using the Matlab software (Matlab 2019a, Mathworks) and analyses of the data
from the Delay-discounting task were performed on the R software (R 3.5.0, R Foundation for
Statistical Computing, 2016).
We only analyzed trials in which the monkeys engaged after the Go-signal, i.e. tried to
perform the action, in order to be able to compute response time (RT). In the Delay-
discounting and the Target-detection tasks, engaging after the Go-signal ensured succeeding
due to the simplicity of the action to be performed. In the Force-discounting task and the
Delay-discounting task, preliminary inspection of the neuronal activity did not reveal
differences across monkeys, so the data were pooled by task.
All the following analyses were performed after z-scoring each parameter (Reward,
Delay, Force category, RT, maximum exerted force) and spike count by neuron.
117
Target-detection task and Force-discounting task studies: response time
determination
The force signal was digitized and sampled at 1kHz. During data acquisition, RTs were
evaluated online and defined as the time when the force signal reached the target threshold.
But to avoid the potential influence of the threshold position, RTs were re-evaluated offline
and defined as the time when the force signal started to diverge from baseline, i.e. the
beginning of the squeeze on the grip. In brief, the mean slope of the force signal was computed
in successive 10ms windows (difference in signal value between the end and the beginning of
the time window). We considered that the animal had started pressing the grip when the slope
was above the maximum slope that could be detected in the baseline period outside of
periods of force exertion, in three consecutive time windows.
Target-detection task and Force-discounting task studies: maximum exerted force
determination
To assess the trial-by-trial maximum exerted force in the two force tasks, we first
subtracted the baseline (mean of the signal in a [-500;0ms] epoch before action onset) from
the force signal. We then looked for the maximum in the force signal in a [0;900ms] epoch
after action onset (based on the RT previously determined), since signal inspection revealed
that the presses lasted between 600 and 900ms.
Behavioral analyses pooling sessions by monkey
To analyze the behavior (RT and engagement) per monkey per task, we pooled the
data from all the sessions for each monkey. In order to conserve across-session variations in
RT, RTs were not z-scored by session before pooling the data of all the sessions together.
However, results were similar when z-scoring RTs per session before pooling the sessions. We
then fit a generalized linear model (GLM) per monkey on the pooled data for each combination
of regressors.
Delay-discounting and Force-discounting tasks: Sliding window procedures for
effects of task parameters and behavioral indicators
Spikes were counted in a 200ms window of time, which were moved forward in 25ms
steps, in a [-800;800ms] epoch around action onset. In order to quantify the effects of cost
118
(Force or Delay), Reward and RT around action onset on spike count, and qualify their
dynamics, we then performed, for each 200ms epoch of the sliding window counting of spikes,
a neuron-by-neuron GLM for spike count with trial-by-trial Reward, cost (Force or Delay) and
RT as parameters.
Delay-discounting and Force-discounting tasks: model comparison for pre-action
spike count
To justify including RT in the GLM explaining the pre-action spike count of LC neurons,
we performed a model comparison procedure. We pooled all the trials from all the neurons,
per task (n=75 for Delay-discounting task, n=92 for Force-discounting task). To take into
account differences in baseline firing rate, we did not z-score the spike count per neuron
before pooling the trials of the different neurons (but results were similar when z-scoring rate
per neuron before pooling the neurons). We also did not z-score RTs per sessions to conserve
across-sessions variations in RT, but similar results were obtained when z-scoring RTs per
session before pooling the sessions. After pooling the sessions, we fit three different GLMs to
the spike count of the neurons (using the fitglm function in MATLAB and the bic.glm function
in R): one with only RT as regressor, one with only task parameters as regressors, and one with
all task parameters and RT. We used Bayesian Information Criteria (BICs) to compare the fit of
the models, but similar results were obtained using Akaike’s information criteria (AIC).
RESULTS
In this study, we tested the hypothesis that the activity of LC neurons around the
triggering of the action was related to effort production, i.e. to the dynamic mobilization of
resources necessary to face a challenge. In that frame, the firing of LC neurons should be
related to both cognitive and physical effort associated with producing an action, rather than
the sensory-motor properties of the action. To test this hypothesis, we compared LC activity
around actions involving cognitive effort, physical effort or both, and examined its dynamic
relation with these two types of effort around action onset. We will first present behavioral
data that provide information about effort production across the three tasks, and then
describe the dynamic modulation of LC activity across these 3 tasks.
119
Behavior
We compared behavior in the three tasks, all involving a manual response to a visual
stimulus (Go-signal, a red point turning green). In each of these tasks, monkeys completed
multiple trials across which actions varied in terms of sensory-motor requirements and/or in
terms of reward contingencies. At the beginning of each trial, monkeys received information
about the current task condition using a specific visual cue and had to adjust their behavior
accordingly. We measured two behavioral responses: engagement (whether they attempted
to perform the trial, or not) and Response Time (RT, the interval between the Go-signal and
action onset). Indeed, even if trials were repeated until correct completion, monkeys could
decide to perform the trial or not. Thus, in each condition, the proportion of trials in which
monkeys engaged (engagement rate) reflected their average willingness to work (WTW) given
the expected sensory-motor constraints and reward contingencies. Critically, since trials were
repeated until correct completion, there was no instrumental interest in refusing to perform
any trial. In such task, the optimal behavior would be to perform all trials regardless of the
associated costs and benefits, because refusing one only increased the delay until reward
delivery. But still, monkeys failed to perform the task more often in trials associated with
higher cost and/or lower rewards, which indicates that they could not repress a natural
tendency to disengage in such conditions, even if in that situation if was counterproductive.
Thus, we assumed that in task conditions associated with lower willingness to work, engaging
in the task and performing the action required a higher level of cognitive control to overcome
the stronger tendency to disengage, compared to conditions in which average willingness to
work was higher. Thus, in each task, we used the contrast in WTW across conditions to
evaluate the influence of these conditions on cognitive difficulty, and potentially on the
amount of cognitive effort mobilized to overcome that difficulty. Since all these tasks involved
triggering a response to a visual target, we could also measure RT, i.e. how quickly the animal
responded to the stimulus. In line with previous studies, we assumed that RTs could be
affected both by sensory effects (how difficult it was to identify the target stimulus), motor
effects (how difficult it was to execute the action), and cognitive effects, namely the amount
of cognitive control required to trigger the action in the current condition (here, as a function
of the natural tendency to disengage from the task in the current condition) (Dmochowski and
Norcia, 2015). Critically, the increase in RT in conditions associated with greater cognitive
control could reflect both the difficulty itself and/or the mobilization of resources (i.e. time)
120
in order to overcome that difficulty (Lee and Daunizeau, 2021; Navon, 1984; Shenhav et al.,
2017). To capture the potential influence of sensory-motor and cognitive effects on RT, we
compared behavior across three tasks manipulating both sensory-motor and reward
parameters across conditions (Fig. 1). The different features of the three tasks are summarized
in Table 1, together with the predicted influence of sensory-motor and cognitive effects on
RT.
Table 1: summary of the characteristics of the three tasks used in this study
Delay-
Discounting
Force-
Discounting
Target-
Detection
Target Detection ✓ ✓ ✓
Var. Position (target & response) 0 0 ✓
Force Production 0 ✓ ✓
Force Constraint 0 ✓ 0
Reward Delay ✓ 0 0
Reward Size ✓ ✓ 0
Sensory motor constraints: 0 ✓ ✓
Cognitive constraints ✓ ✓ 0
In short, all three tasks involved detecting a simple visual target (a red dot turning
green). In the Delay-discounting and in the Force-discounting task, the target stimulus was
always presented in the middle of the screen and only one responding device was available.
In the Target Detection task, the target stimulus could be presented in one of nine possible
positions on the screen and there were three response devices (left, middle and right). In the
Delay-discounting task, the action was a simple bar release whereas in Force-discounting and
Target-detection task, monkeys had to squeeze a grip, and exert a given level of force. In the
Force-discounting task, the level of force necessary to complete the trial was varied
121
systematically across trials according to three difficulty levels, whereas in the Target-detection
task, the level of force required was set to a minimum and equivalent across all task
conditions. Finally, we also manipulated Reward parameters: In both the Delay-discounting
and Force-discounting tasks, we systematically varied the size of the reward (volume of juice)
according to three levels. In the Delay-discounting task, we also systematically varied the delay
between correct responses and reward delivery, according to three levels.
Based on these features, we predicted that, across task conditions, WTW should differ
only in Delay-discounting and Force-discounting tasks, but not in the Target-detection task.
Accordingly, RTs should be affected only by cognitive effects in the Delay-discounting task. In
the Target-detection task, however, RTs should only be affected by sensory-motor effects.
Finally, in the Force-Discounting task RTs should display both sensory-motor and cognitive
effects. To validate these predictions, we compared engagement rates and RT modulations
across conditions in the three tasks.
In the Delay-discounting task, there were clear differences in WTW across conditions:
engagement rates showed a significant positive modulation by Reward and a significant
negative modulation by Delay in both monkeys. For each monkey, we fit a logistic regression
for WTW with Reward and Delay as parameters. The Reward effect was significantly positive
for both monkeys (Monkey T: beta=0.40; p<10-11; Monkey L: beta=0.23; p<10-4) and the
Delay effect was significantly negative for both monkeys (Monkey T: beta=-0.42; p<10-14;
Monkey L: beta=-0.31; p<10-8, Fig. 2A&B). Since in this task sensory-motor constraints were
equivalent across conditions, we expected task conditions to affect RTs through cognitive
effects, and thus according to their influence on WTW. Indeed, in both monkeys, task
parameters had significant opposite effects on RT, as measured using linear modeling (GLM,
Fig. 2C&D). Reward had a negative effect on RT (Monkey T: p<10-34; Monkey L: p<10-33) and
Delay had a positive effect (Monkey T: p<10-95; Monkey L: p<10-52). Thus, task parameters
clearly affected WTW and consequently the cognitive difficulty associated with engaging in
the task and triggering the action. The corresponding differences in RTs across conditions
confirmed this interpretation in terms of cognitive difficulty to trigger the actions, but also in
terms of resources (time, at least) invested in order to overcome the difficulty and perform
the action.
122
Figure 2: Behavior in the Delay-discounting task, the Target-detection task and the Force-
discounting task.
123
(A-D) Behavior in the Delay-discounting task.
(A,B) Coefficients of the logistic regression for engagement in trials (0 if the monkey made no
action, 1 if he performed an action) with task parameters (Reward and Delay) as regressors,
by monkey. For both monkeys, Reward had a positive effect and Delay had a negative effect
on the probability to engage in trials.
(C,D) Coefficients for the GLM for RT with task parameters (Reward and Delay) as regressors,
by monkey. For both monkeys, Reward had a negative effect and Delay had a positive effect.
(E-H) Behavior in the Target-detection task.
(E) Mean engagement rate across sessions by dot position on screen. There was no difference
in engagement rate across the nine conditions (ANOVA).
(F) Mean RT across sessions by dot position on screen. A two way-ANOVA with vertical and
horizontal dot coordinates showed that RTs were significantly shorter for middle grip presses.
(G) Time course of the exerted force on each grip across sessions after action onset. Thick lines
represent the mean exerted force and the thinner lines represent one the standard error of the
mean (SEM) above and below the mean. The maximum exerted force was higher for middle
grip presses (blue) than for right grip (green) or left grip (red) presses.
(H) Coefficients for the GLM for RT with only the maximum exerted force as parameter. RT was
longer if the maximum exerted force would be stronger.
(I-N) Behavior in the Force-discounting task.
(I,J) Coefficients of the logistic regression for engagement in trials (0 if the monkey made no
action, 1 if he performed an action) with task parameters (Reward and Force) as regressors,
by monkey. For both monkeys, Reward had a positive effect and Force had a negative effect
on the probability to engage in trials.
(K,L) Coefficients for the GLM for RT with task parameters (Reward and Force) as regressors,
by monkey. For both monkeys, Reward had a positive effect (only a tendency for monkey D)
and Force had a negative effect.
(M,N) Coefficients for the GLM for RT with only the maximum exerted force as parameter. For
both monkeys, there was a strong negative relationship between RT and the maximum exerted
force.
**: p<0.01; ***: p<0.0001; n.s.: non-significant, error bars represent standard error of the
mean (SEM); RT: response time
124
In the Target Detection task, sensory-motor constraints did vary across task conditions
(positions of targets and of response levers) but they did not affect WTW (one-way ANOVA on
session-by-session rates of engagement in each condition, dot position as parameter, p>>0.05,
Fig. 2E). By contrast, these sensory-motor constraints did affect RTs: we fit a two-way ANOVA
for RT with the horizontal position of the dot (i.e. position of the grip to be used: left middle,
right), and vertical position of the dot as parameters (low, intermediate, high). The effect of
the grip used was significant (F(2)=27.06; p<10-4, Fig. 2F), with lower RTs when using the
middle grip (post-hoc t test with correction for multiple comparison, p<0.05), but no effect of
the vertical position of the dot (p>>0.05). These constraints were also associated with
differences in terms of force production: we fitted an ANOVA for the exerted force with the
side of the grip as a parameter. The effect of side was significant (F(2)=858.6 ; p<10-20) with
the force applied on the middle grip being the highest (multiple comparison of means, p<0.01,
Fig. 2G). We finally confirmed the link between RT and exerted force by looking at the
relationship between RT and the peak of exerted force by fitting RT and exerted force with a
linear model. The effect of the exerted force on RT was significant and negative (p<10-14, Fig.
2H). In summary, in the Target Detection task, the absence of WTW contrast across conditions
indicated that the amount of cognitive control required to trigger the action (press) was
probably equivalent across conditions. Additionally, RTs were clearly modulated by sensory-
motor constraints but there is no reason to interpret differences in RT in terms of cognitive
difficulty and/or associated cognitive effort.
In the Force-discounting task, as described in the original paper (Varazzani et al., 2015),
WTW was strongly modulated across conditions. Using a logistic regression, we found that
engagement was modulated positively by Reward and negatively by Force category (p<0.01
for both parameters for both monkeys, Fig.2I&J). Thus, as in the Delay-discounting task, the
amount of cognitive control required to perform the task should be modulated across task
conditions, and RTs should be influenced by cognitive effects. We then examined the
modulations of RTs across task conditions by fitting a GLM for RTs, with the offered Reward
and the category of Force requested as parameters. In both monkeys, RTs were positively
modulated by Reward (Monkey A: beta=0.58; p<10-4; Monkey D: beta=0.00076; p=0.47) and
negatively modulated by Force category (Monkey A: beta=-0.13; p<10-20; Monkey D: beta=-
0.087, p<10-15), which is the opposite of the pattern expected for pure cognitive effects, given
the influence of task parameters on WTW (Fig.2K&L). But these effects of task parameters
125
could also be accounted for in terms of simple motor constraints, rather than cognitive
constraints. To explore that possibility, we examined the relationship between trial-by-trial
RTs and exerted force (maximum exerted force on the grip during press). As was the case in
the Target-detection task, there was a significant negative relation between RT and the
amount of exerted force (GLM, p<0.01 for both monkeys, Fig. 2M&N). Thus, in this task, RTs
were clearly affected by sensory-motor constraints. However, even if the pattern of WTW
suggests that they might also be affected by cognitive constraints, we could not find evidence
for it.
In short, even if all three tasks involved triggering an action in response to a visual
target, the three tasks differed in the relative weight of sensory-motor versus cognitive
constraints on RT. In the Delay-Discounting task, responses were mostly influenced by
cognitive constraints, with virtually no difference in sensory-motor processes across
conditions. By contrast, in the Target-Detection task, conditions clearly differed in terms of
sensory-motor constraint but not in terms of value. In the Force Discounting task, the
difference in behavior across conditions suggests that it involved both sensory-motor and
cognitive constraints, but the relative weight of the two remains difficult to evaluate.
Neurophysiology:
We recorded single LC units based on previously described methodologies (see
materials and methods for details). In the Delay-discounting task, we recorded the activity of
75 single units from the LC (n=52 in Monkey T and n=23 in Monkey L). In the Force-discounting
task, we recorded 92 LC units (in 2 monkeys, n=63 in Monkey D and n=29 in monkey A). In the
Target-detection task, we recorded 26 neurons (in one monkey, Monkey J). The baseline firing
rate of all the recorded neurons remained relatively constant during the course of the
experiments, besides the periods when monkeys were apparently asleep (eyes closed and
motionless) and LC neurons were virtually silent. Within tasks, inspection of the data did not
indicate any difference between neurons recorded from the different monkeys, so the
neurons were pooled by task.
Evoked responses to action onset and action execution
As classically described in previous studies, LC neurons were activated just before
action onset (Bouret and Richmond, 2015, 2009; Bouret and Sara, 2004; Rajkowski et al.,
126
2004). Neuron-by-neuron, we measured firing rates in a [-250;0ms] epoch before action onset
and also in a baseline period, just before the onset of the Go-signal ([-250;0ms] from Go-
signal). We compared these rates using paired t-tests and then used a second-order test (t-
test on T values) to assess the coherence of the changes in rate. Rate significantly differed
from baseline for 62/75 neurons in the Delay-discounting task (all increased, Fig. 3A), 51/92
neurons in the Force-discounting task (all increased, Fig. 3B); and 16/26 neurons in the Target-
detection task (8 decreased and 8 increased, Fig. 3C). At the population level, the activation
only reached significance in the Delay- and Force-discounting tasks (second level analysis:
Delay-discounting: (t(74)=9.56; p<0.001, Fig. 3D; Force-discounting: t(91)=8.88; p<10-13, Fig.
3E). In the Target-detection task, the change in rate did not reach significance at the
population level (t(25)=0.57; p=0.58, Fig. 3F).
In addition to the activation right before action onset, LC neurons were also activated
during the action itself in both tasks involving force production (Force-discounting and Target-
detection). In the Force-discounting task, in the [0;600ms] epoch after action onset,
corresponding to the time during which the force was exerted (Fig. S1), 50/92 neurons showed
a significant change in rate compared to pre-go signal rate (10 decreased and 40 increased,
Fig. 3B) and the activation was clearly significant at the population level (t(91)=15.29; p<10-26,
Fig. 3E). In the Target-detection task, in the [0;500ms] after action onset during which force
was exerted (Fig. 2G), 20/26 neurons showed a significant change in rate during force
production (6 decreased and 14 increased) and the activation was significant at the population
level (t(25)=7.38; p<10-7, Fig. 3F). On the contrary, rate decreased after action onset in the
Delay-discounting task, compared to the baseline rate. The change was significant for 36/75
neurons (all decreased, t(74)=-2.14; p=0.035; Fig. 3A&D). In that task, the action was much
shorter and less demanding (a small hand movement).
According to our hypothesis that LC neurons encode instantaneous effort production,
the activation of LC neurons before action should reflect the amount of cognitive effort
associated with the decision to produce the action given the natural tendency to engage or
not in the trial. By contrast, the activation of LC neurons during force production should reflect
the amount of physical effort involved in the action. To test this hypothesis, we compared
firing rate modulation around action onset across conditions and across the three tasks.
127
Figure 3: Evoked responses to action onset in the Delay-discounting, Force-discounting and
Target-detection task.
(A-C) Mean rate (colors) of each of the recorded units (y-axis) around the time of action onset
(x-axis) in (A) the Delay-discounting task, (B) the Force-discounting task, and (C) the signal-
detection task. Clearer colors represent higher rates.
(D-F) Results of the first-order and second-order t-tests on rate in (D) the Delay-discounting
task, (E) the Force-discounting task and (F) the Target-detection task. Density function of the
T-values of the neuron-by-neuron t-tests comparing rate before action onset ([-250;0ms] from
action onset, yellow), during action execution ([0;600ms] from action onset, dark red) or after
action onset ([0;250ms] after action onset, green) to baseline rate (rate before go-signal, [-
250;0ms]).
(D) In the Delay-discounting task, across the population, rate increased before action onset
and decreased after action onset.
(E) In the Force-discounting task, across the population, rate was above baseline before action
onset and during action execution.
128
(F) In the Target-detection task, across the population, rate was not different from baseline
before action onset and was higher than baseline during action execution.
*: p<0.05; ***: p<0.001; n.s.: non-significant.
Modulation of action-onset related activity by cognitive and physical effort
To evaluate the modulation of LC activity associated with the triggering of the action
based on the associated cognitive difficulty and effort, we examined the modulation of firing
rates just before action onset across task conditions. We also examined the influence of RTs,
which, in absence of sensory-motor constraints, capture trial-by-trial modulations of cognitive
difficulty and/or the associated cognitive effort to trigger the response (see above, ‘Behavior’).
To assess the potential relation between LC and physical effort, we computed spike
counts in a [0;600ms] epoch after action onset, based on force profiles in both tasks. We
compared LC activity across conditions both in the Force-discounting task (where the force
was imposed and represented an effort, as measured using WTW) and in the Target-detection
task (where we could not find any evidence of cognitive modulation since WTW was
equivalent across conditions). We also examined the relation between firing rates and force
production, with the hypothesis that the amount of force produced only reflected effort in the
Force-discounting task, but not in the Target-detection task.
LC and cognitive effort: Action triggering-related activity
Delay-discounting task
To examine the influence of task parameters on action triggering-related activity in the
Delay-discounting task in each neuron, we fit spike counts right before action onset with a
GLM using Reward and Delay as parameters. The activity of an example unit encoding task
parameters is shown in Figure 4A&B. 9/75 LC units significantly encoded Reward, negatively.
Even if that number was close to the number of neurons expected by chance (n=10) given a
population of that size (n=75), the negative effect of Reward was significant across the
population (second order t-test: t(74)=-4.18; p<0.001, Fig. 4C). There was also a significant
positive effect of Delay for 20/75 neurons, and this positive effect of Delay was consistent
across the population (t(75)=5.44; p<0.001, Fig. 4C). Thus, when monkeys triggered the action,
the direction of the effects of task parameters on LC activity was opposite to what we reported
for WTW (Fig. 2I&J).
129
Next, we examined the influence of RT on action triggering-related activity by fitting a
GLM. The activity of an example unit encoding RT positively is shown in Figure 4D. The firing
of 12/75 units exhibited a significant effect of RT (positive for 11 of them) and the positive
influence of RT was significant across the population (second-order t test: t(74)=4.18; p<0.001,
Fig. 4E). When task parameters were included as covariates of RT in the regression model,
8/75 neurons still exhibited a significant positive effect of RT. Critically, the positive influence
of RT was still significant at the population level (t(74)=2.21; p=0.024, Fig. 4F). Note that both
Reward and Delay still had significant and opposite effects on LC activity (Reward, significant
for 8/75 neurons, 7 negative and 1 positive, second order test: t(74)=-3.74, p=0.0003 ; Delay:
significant for 17/75 neurons, all positive, second order test: t(74)=5.05, p<10-5).
Finally, we compared the fit of the GLMs for spike count including RT alone, task
parameters and a full model with both RT and task parameters. We fit each of these three
GLMs on the entire dataset (all trials of all sessions of both monkeys, spike count not
normalized by neuron) and compared the Bayesian Information Criteria (BIC) of those three
models. The BIC of the model with Reward, Delay and RT was lowest (BIC=35742), compared
to the BIC of the model with only RT (BIC= 35874) and to the BIC with only task parameters
(BIC=35760). Thus, modulation of LC activity across conditions could be accounted for by both
task parameters, which are related to the difficulty to make the decision (to perform that
action), and to the RT, which increases with the difficulty to trigger the action and reflects the
amount of resources (at least time) invested in triggering the action.
Force-Discounting task
We assessed the encoding of task parameters by LC neurons using a GLM for spike
count before action onset with Reward and Force category as parameters. The activity of
example LC units encoding task parameters is shown in Figure 4G&H. The effect of Reward
was only significant for 6/92 neurons (1 positive, 5 negative), which is less than the number of
neurons expected by chance (n=13) for a sample of 92 neurons. At the population level,
however, the effect of Reward was significantly negative (second order t-test: t(91)=-2.05;
p=0.04, Fig. 4I). Similarly, the effect of Force was only significant for 7/92 neurons (1 negative,
6 positive) but it was significantly positive at the level of the population (t(91)=2.72; p=0.0079,
Fig. 4I). Thus, the influence of task parameters on LC activity at time of action triggering
mirrors their influence on WTW.
131
Figure 4: Encoding of proxies for cognitive effort to trigger the action and physical effort to
execute it, before action onset in (A-F) the Delay-discounting task, (G-M) the Force-discounting
task and (N) the Target-detection task.
(A,B) Raster plots of the activity of two representative LC units encoding task parameters
(Delay, (A) and Reward, (B)) recorded in the Delay-discounting task, around action onset
(orange vertical line). Green dots represent the go-signal. Trials are split by (A) Delay levels
(D1,2,3) or (B) Reward size (R1,2,3). Unit (A) encoded Delay positively before action onset, i.e.
was more active before action onset in high-Delay conditions. Unit (B) encoded Reward
negatively, i.e. was less active before action onset in high-Reward conditions.
(C) Summary (mean and SEM) of the coefficients of the neuron-by-neuron GLMs for spike count
before action onset with Reward and Delay as parameters, in the Delay-discounting task.
Across the population, neurons encoded Reward negatively and Delay positively.
(D) Raster plot of the activity of a representative LC unit encoding RT positively around action
onset (orange line) in the Delay-discounting task. This unit was more active when RT (interval
between Go-signal, green dot, and action onset) was longer.
(E,F) Summary (mean and SEM) of the coefficients of the neuron-by-neuron GLMs for spike
count before action onset with (E) RT alone or (F) RT, Reward and Delay as parameters in the
Delay-discounting task. (E) Across the population, neurons encoded RT positively, (F) and this
positive encoding was true over and above the encoding of Reward (negatively) and Delay
(positively).
(G-H) Raster plots of the activity of two representative LC units encoding task parameters
(Force category , (G) and Reward, (H)) recorded in the Force-discounting task, around action
onset (orange vertical line). Green dots represent the go-signal. Trials are split by (G) Force
levels (F1,2,3) or (H) Reward size (R1,2,3). Unit (G) encoded Force positively before action
onset, i.e. was more active before action onset in high-Force conditions. Unit (H) encoded
Reward negatively, i.e. was less active before action onset in high-Reward conditions.
(I) Summary (mean and SEM) of the coefficients of the neuron-by-neuron GLMs for spike count
before action onset with Reward and Force category as parameters, in the Force-discounting
task. Across the population, neurons encoded Reward negatively and Force category positively.
(J) Raster plot of the activity of a representative LC unit encoding RT positively around action
onset (orange line) in the Force-discounting task. This unit was more active when RT (interval
between Go-signal, green dot, and action onset) was longer.
132
(K,L) Summary (mean and SEM) of the coefficients of the neuron-by-neuron GLMs for spike
count before action onset with (K) RT alone or (L) RT, Reward and Force category as
parameters, in the Force-discounting task. (K) Across the population, neurons encoded RT
positively, (L) and this positive encoding was true over and above the encoding of Reward
(negatively) and Force category (positively).
(M,N) Summary (mean and SEM) of the coefficients of the neuron-by-neuron GLMs for spike
count before action onset with (M) the maximum exerted force on the grip alone or (N) the
maximum exerted force on the grip, Reward and Force category as parameters, in the Force-
discounting task. (M) Across the population, neurons marginally encoded the maximum
exerted force, (N) but this effect was not present over and above effects of task parameters
(Force category and Reward).
(O) Summary (mean and SEM) of the coefficients of the neuron-by-neuron GLMs for spike count
before action onset, with RT as parameter, in the Target-detection task. Across the population,
neurons did not encode RT.
*: p<0.05; **: p<0.01; ***: p<0.001; n.s.: non-significant; D: Delay, F: Force, R: Reward; RT:
response time.
We examined the encoding of RT by LC neurons by fitting a GLM for spike count with
RT alone (in addition to the constant term). The encoding of RT by an example unit is shown
in Figure 4J. The effect of RT was only significant for 10/92 neurons (4 negative, 6 positive) but
it was consistently positive at the level of the population (t(91)=2.01; p=0.048, Fig. 4J). To
evaluate the possibility that the relation between spike counts and RT was confounded by a
joint relation with task parameters, we added task parameters (Reward and Force category)
as co-regressors, along with RT. The effect of RT was still significant for the same 10/92
neurons (7 positive and 3 negative) and it remained consistently positive across the population
(t(91)=2.38; p=0.020, Fig. 4K). Thus, LC neurons encoded RT just prior action onset, over and
above task parameters. Note that the effect of Force category was still significant for 6/92
neurons (5 positive, 1 negative) and was consistently positive across the population
(t(91)=2.86; p=0.0052, Fig. 4K). The effect of Reward was still significant for 5/92 neurons (4
negative, 1 positive) and second order statistics confirmed that it remained significant at the
population level, with a significant negative effect (t(91)=-1.88; p=0.048, Fig. 4K).
133
We next examined the relation between LC activity before action onset and the
amount of force exerted during action itself, using a GLM with only the maximum exerted
force as parameter (besides the constant term). The effect was significant for 14/92 neurons
(2 negative, 12 positive) and marginally significant at the population level (second order t-test:
t(91)=1.68; p=0.096, Fig. 4L). When Force category and Reward were included as co-regressors
in the model, the influence of the exerted force on firing rate remained significant for 10/92
neurons (5 positive, 5 negative), but it showed no significant effect at the population level
(second-order t-test: p=0.95, Fig. 4M). Thus, the influence of the amount of force exerted on
the grip during the action on the activity of LC neurons just before action onset was negligible,
compared to that of RT.
Finally, we compared the fit of the GLMs for spike count including RT, the exerted
force, task parameters and combinations of those. To run this model comparison on the entire
population of neurons, as we did for the Delay-discounting task, we pooled all the trials of all
the sessions into a single database. Then we fit five GLMs on the data: one with only RT as
parameter, one with Reward and Force category as parameters, and one with RT, Force
category and Reward, one with the maximum exerted force alone, and one with Force
category, Reward and the maximum exerted force. The BIC of the model with RT alone was
lowest (BIC=54960), compared to the BIC of the model with only task parameters (BIC=54972),
with RT and task parameters (BIC= 54969), with only the peak of force (BIC= 54976), and with
the peak of force and task parameters (BIC=54983). Thus, the activity of LC neurons before
action onset in this task was more strongly influenced by RT than by the amount of force about
to be exerted on the grip.
Target-detection task
In the Delay-discounting and Force-discounting tasks, we found a positive encoding of
RT prior to action onset. In both of these tasks, behavioral analyses indicates that triggering
the action involved distinct levels of cognitive difficulty (and potentially cognitive effort) and
trial-by-trial RT could be taken as a proxy for subjective difficulty and corresponding effort.
But since RT modulations across conditions also included sensory-motor constraints,
especially in the Force-discounting task, we examined LC activity in the Target-detection task
as a control for these effects. Indeed, in the Target-detection task, we could not detect any
evidence for a systematic modulation of subjective difficulty/cognitive effort across conditions
134
but RTs showed clear modulations, that could be interpreted in terms of sensory-motor
constraints (see above ‘Behavior’). To examine the potential modulation of LC activity, we fit
a GLM for spike count in the [-250;0ms] epoch before action onset in the Target-detection
task, with only RT as parameter. Only 1/26 neuron exhibited a significant (positive) encoding
of RT. Moreover, the influence of RT on spike count was not significant at the population level
(p=0.56, Fig. 4N).
Thus, in this task where the behavioral response was modulated by sensory-motor
processes, LC neurons showed no significant modulation of action triggering-related activity
across task conditions and no relation with the RT.
LC and physical effort: force exertion-related activity
To examine the relation between LC activity and force production in the Force-
discounting and Target-detection tasks, we focused on the period of the force exertion, after
action onset. Force exertion usually lasted for about 800ms in the Force-discounting task (Fig.
S1A&B), and 500ms in the Target-detection task (Fig. 2G), with the shortest presses lasting for
about 600ms and 400ms respectively. In order to avoid any overlap with the reward delivery
period, we restricted the analyses to a [0;600ms] post action onset epoch in the Force-
discounting task and a [0;400ms] epoch in the Target-detection task.
Force-discounting task
We first examined the modulation of LC activity by task parameters by fitting a GLM
for spike count after action onset (during force exertion) with Reward and Force category. The
effect of Reward was only significant for 11/92 neurons (3 negative, 8 positive), which is less
than the numbers of neurons expected by chance (n=13) given the size of that population
(n=92). In addition, the influence of Reward was not significant across the population (second
order t-test, p=0.7, Fig. 5A). The effect of the Force category was significant for 38/92 neurons
(1 negative, 37 positive, example unit in Fig. 4G) and was consistently positive across the
population (t(91)=9.43; p<10-14, Fig. 5A).
We then examined the influence of physical force by fitting a GLM for spike count with
the maximum exerted force as parameter. The effect was significant for 38/92 neurons (2
negative, 36 positive) and it was consistently positive across the population (second order t-
test, t(91)=7.39; p<10-10, Fig. 5B). In addition, the influence of exerted force on LC activity
135
during action execution remained significant when task parameters were included as co-
regressors (exerted force effect: still significant for 17/92 neurons, 4 negative, 13 positive;
population effect: t(91)=2.04 ; p=0.044, Fig. 5C). Note that the effect of Reward was significant
for 10/92 neurons (3 negative, 7 positive), and still not significant at the level of the population
(second order t-test: p=0.66, Fig. 5C). The effect of the category of Force was significant for
20/92 neurons (3 negative, 17 positive) and it remained significantly positive at the population
level (second order t-test: t(91)=4.08; p<10-4, Fig. 5C). Thus, the activity of LC neurons during
force production was significantly modulated by the amount of force exerted, over and above
the influence of task parameters.
We then examined the influence of RT on LC activity during force exertion. The effect
of RT was significant at the population level when RT was the only parameter in the GLM for
spike count (significant for 10/92 neurons, 3 positive, 7 negative; second order t-test: t(91)=-
2.94; p=0.0041). However, the effect was no longer significant when task parameters were
added to the model (still significant for 10/92 neurons, second order t-test: p=0.18), indicating
that during force exertion, LC neurons showed little sensitivity to RT.
Target-detection task
We examined influence of the exerted force on LC activity during force exertion in the
Target-detection task. We fit a GLM for spike count in the [0;400ms] post-action onset epoch
with only the maximum exerted force as parameter. The effect was significant for 7/26
neurons (3 negative, 4 positive) but it was not significant across the population (p=0.62, Fig.
5F). Note that the effect of the exerted force was not present either before action onset
(significant for 4/20 neurons, 3 positive, 1 negative, second order t-test: p=0.56), and that
there was no effect of RT on spike count during force exertion (significant for 3/26 neurons, 2
negative, 1 positive, second order t-test: p=0.3, Fig. 5G).
Thus, in this task where there is no evidence that force production involved a
systematic difference in effort across conditions, we found no evidence for a modulation of
LC activity by amount of force produced.
136
Figure 5: Force-exertion-related activity in the Force-discounting and Target-detection tasks.
(A-E) Summary (mean and SEM) of the coefficients of the neuron-by-neuron GLMs for spike
count during action (force exertion) with (A) task parameters (Force category and Reward) or
(B) the maximum exerted force on the grip alone, (C) task parameters and the maximum
exerted force, (D) RT alone, or (E) RT, Reward and Force category as parameters, in the Force-
discounting task.
(A) During the force exertion, LC neurons encoded the Force category positively if the exerted
force was not included in the model, and (B) they encoded the exerted force positively if task
parameters were not included in the model. (C) In the model including task parameters and
the maximum force exerted, there was a positive effect of both the exerted force and Force
category.
(D) During the force exertion, there was a positive effect of RT on spike count if RT was alone
in the model, but (E) this effect disappeared if task parameters were included in the model.
137
(F,G) Summary (mean and SEM) of the coefficients of the neuron-by-neuron GLMs for spike
count during action (force exertion) with (F) the maximum exerted force on the grip alone, or
(G) RT alone, as parameters, in the target-detection task. (F) During force exertion, there was
no effect of the maximum exerted force (G) or of RT in the Target-detection task.
*: p<0.05; ***: p<0.001; n.s.: non-significant; F: Force, R: Reward; RT: response time.
DISCUSSION
Summary
The aim of this study was to clarify the dynamic relation between LC activity and effort
production. We examined the activity of LC neurons in two effort-discounting tasks (Force-
discounting and Delay-discounting) and compared it to activity in a task only manipulating
sensory-motor processes but not effort (target-detection task). First, in both effort-
discounting tasks, triggering the response required more producing cognitive control when
outcome value decreased. The firing of LC neurons at the time of action-triggering scaled with
task parameters in directions compatible with an encoding of the difficulty of triggering the
action, and with RT, which probably scaled with the effort produced to trigger the action.
Second, LC neurons were activated during force production both in the Force-discounting task
and in the control Target-detection task, but the magnitude of this activation only scaled with
the exerted force in the Force-discounting task, where producing the force was actually
effortful (since it decreased trial subjective value, as measured by WTW). In the control task,
where similar sensory-motor processes where involved but where we could not find any
evidence of an effort contrast across conditions, as evidenced by the absence of engagement
differences, LC neurons showed no sign of modulation by the exerted force. Thus, across tasks,
LC neurons were activated at the time when monkeys made an effort, either a cognitive effort,
to trigger a costly behavioral response, or a physical effort, to exert force on the grip. Critically,
the magnitude of these activations scaled with the effort produced to face that difficulty, with
a sub-second precision. Altogether, this work supports the idea that LC activity continuously
monitors the production of effort.
138
Behavior
Manipulating effort in animals is very challenging, since it is a complex notion and
measures remain very indirect. Operationally, effort can generally be defined as a mobilization
of resources to boost performance in the face of a difficulty and meet task goals (Borderies et
al., 2020; Kool and Botvinick, 2018; Shenhav et al., 2017). Therefore, effort can only be
evaluated indirectly, through the relative modulations of performance relative to task
difficulty. Here, we manipulated two types of difficulty: overcoming a natural tendency to
disengage from the task when the expected value was low (cognitive effort), and exerting a
relatively high level of force on the grip (physical effort).
For cognitive effort, we took advantage of the fact that monkeys tend to disengage
from single-option tasks when the outcome value is low, even if erroneous trials are repeated
(Bowman et al., 1996; Hori et al., 2009; Jahn et al., 2018, 2020; Minamimoto et al., 2009b,
2012; Varazzani et al., 2015). In these conditions, failing to engage in a low-value trial carries
disutility, since it only increases the delay until the next reward: from an instrumental
perspective, monkeys should engage in every single trial until their motivation level is low
enough for them to stop performing the task. Conversely, in versions of these tasks where
aborted trials were not repeated, monkeys rapidly learned to skip trials associated with low
value and/or high costs (La Camera et al., 2018). This tendency to abort trials associated with
low benefits and/or high costs implies a greater difficulty to overcome this automatic response
and perform the task despite low short-term utility. When erroneous trials are repeated and
skipping trials is not instrumental, including our tasks, monkeys engage in a majority of trials.
Consequently, in correctly performed trials in our tasks, they must have been exerting
cognitive control to overcome their natural tendency to abort trials associated with low
reward and/or high costs, as classically described in humans and other animals performing
standard discounting tasks (Filevich et al., 2013, 2012; Shenhav et al., 2017; Widge et al.,
2019), although the nature and the dynamics of the underlying mechanisms remains difficult
to assess. Based on that principle, we assumed that the proportion of trials in which monkeys
failed to engage in a specific task condition was a good proxy of the effort required to trigger
the action (subjective difficulty) in that cost-benefit contingency (i.e. the expected outcome
value). In both the Delay-discounting and Force-discounting tasks, the subjective difficulty to
trigger the action (and presumably the associated level of cognitive effort) decreased with
expected benefits (reward size) and increased with expected costs (force or delay). By
139
contrast, we failed to detect any evidence of a difference in WTW across conditions in the
control Target-detection task, which implies that in that task the level of subjective difficulty
and associated cognitive control invested to engage in the task was relatively equivalent
across conditions.
In addition, we used RTs as a proxy for trial-by-trial variability in subjective difficulty to
trigger the action, and potentially to the corresponding amount of cognitive effort expensed.
Obviously, RTs were affected both by decision-related processes and by sensory-motor
constraints (Dmochowski and Norcia, 2015; Gilden, 1997; Hunt et al., 2012; Krajbich et al.,
2010; Krajbich and Rangel, 2011; Milosavljevic et al., 2010; Morris et al., 2019; Philiastides and
Ratcliff, 2013). Still, in tasks in which conditions differed almost exclusively in terms of
outcome value, with virtually no sensory-motor differences, RTs reliably decreased with
outcome value and therefore increased with the subjective difficulty to trigger the response
(Hori et al., 2009; Minamimoto et al., 2009b, 2012). In both of our tasks involving force
production (Force-discounting and Target-detection), RTs were strongly modulated by motor
processes, as indicated by the significant relation with the amount of exerted force. But in the
Delay-discounting task, in which sensory-motor constraints were equivalent across conditions,
the modulations of RTs were in line with an interpretation in terms of subjective difficulty to
trigger the action (to overcome a natural tendency to abort trials when value was low). Since
monkeys remained engaged in the task from cue onset to outcome delivery (especially in the
Force-discounting task in which gaze fixation was imposed), variations of RTs across trials were
equivalent to measures of time on task. Under the assumption of a constant rate of effort
production per unit of time, time on task and therefore RT could be taken as a measure of the
amount of cognitive effort invested in the decision to trigger the action (Griffiths et al., 2015;
Lieder et al., 2014, 2012; Lieder and Griffiths, 2017). On the other hand, one could argue that
for a given level of difficulty (within task conditions), engaging more effort would result in
making decision processes more efficient, and therefore lead to shorter RTs (Navon, 1984;
Shenhav et al., 2017).
In humans studies, pupil dilation often appears to constitute a better measure of
cognitive effort, compared to RT (Alnæs et al., 2014; Hess and Polt, 1964; Kahneman, 1973;
Kahneman and Beatty, 1966; Simpson and Hale, 1969; Wahn et al., 2016; Wright et al., 2013).
But given the complex and slow dynamics of pupil dilation it would have been difficult to use
it here, especially given the strong relation between pupil and physical effort (Varazzani et al.,
140
2015; Zénon et al., 2014). From a theoretical point of view, the strong correlation between an
autonomic measures and cognitive and physical effort suggests that the central nervous
system could be using a common mechanism to mobilize the resources in the face of both
cognitive and physical challenges. In both cases, the activation of the sympathetic system
would not only increase pupil dilation but also heart rate and blood pressure, as well as the
availability of metabolic resources. From a practical point of view, this also implies that pupil
is not a specific measure of cognitive effort. Finally, the use of pupil dilation as a measure of
effort is often interpreted in the light of its strong relation with LC activity (Joshi et al., 2016;
Murphy et al., 2014; Varazzani et al., 2015). But obviously, since the goal of these experiments
was precisely to evaluate the actual relation between effort and LC activity, the interest of
pupilometry would have been limited.
Thus, for the purpose of this study, we feel relatively confident that trial-by-trial
variations of RT provide a good proxy for the subjective difficulty to trigger the responses, and
we can only speculate that they were also related to the amount of cognitive effort associated
with facing the difficulty. Of course, further experiments would be necessary to clarify that
point, since it is unclear how increasing the resources allocated for a decision process could
cause a slowing of responses rather than a speeding (Navon, 1984; Shenhav et al., 2017). Still,
we feel that demonstrating a strong and specific relation with LC activity at the time of the
decision to trigger the action provides a strong basis for such research plans.
LC activity
In both our effort-discounting tasks, LC neurons activated just prior to action onset.
Importantly, such activation was never seen for actions executed outside of tasks (between
trials for example), indicating that it was probably associated to the decision process of
triggering actions rather than movements per se. In coherence with that, pupil diameter has
been repeatedly shown to increase during decisions (Beatty, 1982; de Gee et al., 2014;
Gilzenrat et al., 2010; Lempert et al., 2015; Nassar et al., 2012), even in the context of covert
decisions (Einhäuser et al., 2010). Noticeably, LC neurons seem to only respond to decisions
to trigger actions and not to withhold them (Kalwani et al., 2014), although this contrast could
merely emerge from the difficulty to identify precisely the timing of decisions that are not
materialized by overt actions. Indeed, since the activations of LC neurons usually consist in a
141
few spikes (sometimes even just one), they could easily be missed if the activity is not aligned
correctly.
In our effort-discounting tasks, the coherent but opposite modulations of LC activity
and engagement rates by task parameters as well as the positive relation between LC activity
and RT provide additional arguments for the relation of LC activity with decision processes,
more specifically effort production for decisions to act. Indeed, task parameters (Reward and
Delay or Force) influenced pre-action onset LC activity and WTW in opposite directions.
Additionally, LC neurons reliably encoded RT in that same period across the two effort-
discounting tasks. In our frame, this modulation points towards an encoding of the subjective
difficulty to trigger actions, and possibly the cognitive control produced, although the strong
motor effects on RT in the force-discounting task called for some caution in that
interpretation. Yet, since we controlled for RT encoding in a context in which it virtually only
varied with sensory-motor effects (Target-detection task), our results point towards a specific
relation between LC activity and subjective difficulty. Additionally, our model comparisons
showed that GLMs including RT predicted LC activity better than models with task parameters
alone, and model fitting confirmed that RT was encoded over and above task parameters in
both effort-discounting tasks. Provided that task parameters represent trial difficulty, this
result would be in favor of an encoding of the allocated effort given the difficulty, or at least
of the invested time.
Similarly, during the sustained actions of the Force-discounting task (lasting a few
hundred milliseconds), LC neurons encoded the force exerted on the grip. Importantly, it was
not the case in a task in which the exerted force had no impact on the tendency to disengage
from trials (WTW in Target-detection task), i.e. in which the force was not perceived as
effortful. Despite the necessary caution in interpreting this negative result, our findings would
be in favor of a specific encoding of physical effort during its exertion. Besides, our
interpretation resonates with the finding that pupil dilation correlates with physical effort,
especially, if not only, in high effort conditions (Zénon et al., 2014).
Overall, our data establishes that LC neurons positively encode the subjective difficulty
of ongoing processes, and possibly the effort allocated, in case of both cognitive control
production and physical effort exertion. In other words, LC neurons as a population exhibit a
domain-general encoding of subjective difficulty and probably effort production.
Complementarily, our recent pharmacological studies are compatible with the idea that LC
142
activation supports the production of effort face of upcoming challenges, at least physical
effort (Borderies et al., 2020; Jahn et al., 2018). Indeed, decreasing NA levels induced an
increase in effort sensitivity and a decrease in force exertion, which could be expected if the
LC/NA system was involved in the mobilization of resources. Conversely, these results are
incompatible with a role of the LC in reporting subjective difficulty and effort costs to effort-
allocating structures. Indeed, in that frame, one would have expected the perturbation of LC
activity to decrease effort sensitivity, as can be seen for reward sensitivity when dopamine
depletions weaken the signaling of rewards (Le Bouc et al., 2016; Martínez-Horta et al., 2014).
We showed that during periods of decreased LC activity, monkeys still processed the effort
dimension of offers but act as if its cost was greater (Borderies et al., 2020). In the same vein,
modifications of brain NA levels and LC activity have been shown to affect cognitive control
production and specifically inhibitory control (Arnsten, 2006; Bari et al., 2020; Chamberlain et
al., 2009; Faraone et al., 2005; Fernando et al., 2012; Navarra et al., 2008; Robinson, 2012;
Robinson et al., 2008), working memory (Chamberlain et al., 2006b; Coradazzi et al., 2016;
Ellis and Nathan, 2001; Gamo et al., 2010; Li et al., 1999; Mao et al., 1999; Tzavara et al., 2006;
Wang et al., 2007), and set-shifting (Lapiz et al., 2007; Lapiz and Morilak, 2006; McGaughy et
al., 2008; Tait et al., 2007), although only a few of these studies used very specific methods.
In light of the current data, it is tempting to suggest that with a decreased NA tone, facing
upcoming challenges is even harder, and vice versa, regardless of the type of challenge
(mental or physical).
The timing of the modulations of LC activity by subjective difficulty and possibly effort
production reported here is striking, as we found that neurons encoded RT very late in the
action-triggering process (at its very end, around 200ms before action onset) and the exerted
force during the execution of the action. LC axons have a very slow conduction velocity due to
their absence of myelin, such that action potentials reach cortical areas in about 130ms
(Aston-Jones et al., 1985). Although we cannot exclude that the dynamics observed in the
present study result from the artificial separation in time of the decisions to engage and the
triggering of the action (demonstrated by Jahn et al., 2020), they would only be compatible
with an a posteriori influence on effort production. In other words, LC activity may promote
effort production following effortful processes, proportionally to their subjective difficulty and
possibly to the effort produced. For instance, the pre-action onset activation may serve to
facilitate action execution while the peri-action activation may enhance action consequence
143
monitoring. In coherence with that idea, some indirect evidence point towards the
involvement of LC activity in action monitoring, and especially post-error adjustments of RTs
(post-error slowing), as pupil diameter around errors predict the strength of post-error
slowing (Murphy et al., 2016) and non-selective enhancements of LC activity enhance it
(Sellaro et al., 2015). LC activity has also been proposed to serve as a network reset signal
(Bouret and Sara, 2005; Corbetta et al., 2008). The timing of modulations observed here could
be compatible with a role in promoting connectivity adaptations at the transition between
processes (e.g. action-triggering to action execution). Additionally, adjustments in brain
connectivity patterns, and more specifically enhancements of connectivity in task-relevant
networks and decreases in task-irrelevant ones (Brink et al., 2016; Guedj et al., 2017b; Hernaus
et al., 2017), could participate in the facilitation of effort production. These ideas remain quite
speculative, and further investigations would be required to clarify the delayed effects of LC
activity on effort production and the underlying mechanisms.
Another unresolved element here concerns the origin of the inputs of the LC that could
provide information about effort production, as well as the targets of the LC responsible for
its effort facilitation effects. Mental and physical effort perception correlates with pupil
diameter (Alnæs et al., 2014; Kahneman, 1973; Wahn et al., 2016; Wright et al., 2013; Zénon
et al., 2014) and is associated to changes in heart rate and metabolic activity (Andrade et al.,
2020; Carroll et al., 1986a; Da Silva et al., 2020; Sims et al., 1988; Sims and Carroll, 1990; Tibana
et al., 2019; Turner et al., 1983; Turner and Carroll, 1985), indicating that effort production is
associated to autonomic activation. The modulations of LC activity by subjective difficulty
could merely be the consequence of its connections with the nucleus Paragigantocellularis
(PGi, Aston-Jones et al., 1991, 1986; Ennis and Aston-Jones, 1988, 1986; Van Bockstaele and
Aston-Jones, 1995), a key component of the sympathetic system. They could also arise from
and/or target cortical and subcortical components of the brain effort system. For instance, in
the thalamus, a subcortical structure widely connected to the LC (“Ascending projections of
the locus coeruleus in the rat. II. Autoradiographic study,” 1977; “Demonstration of extensive
brainstem projections to medial and lateral thalamus and hypothalamus in the rat,” 1990;
McBride and Sutin, 1976; Schwarz et al., 2015), the centromedian-parafascicular complex
(Cm-Pf) exhibits task-evoked activity reminiscent of LC patterns of responses to relevant cues
and actions, with regards to trial value (Matsumoto et al., 2001; Minamimoto et al., 2009a,
2005; Minamimoto and Kimura, 2002). The Cm-Pf, possibly via interplay with the LC, could
144
play a similar role in regulating effort production. Another region considered as a crucial
component of the effort system is the medial pre-frontal cortex, and in particular its
subcomponent the anterior cingulate cortex (ACC), although the anatomical definition of that
area remains quite volatile across studies, and might at least overlap with the supplementary
motor area (SMA). The ACC-SMA is critical for physical effort perception (White et al., 2013;
Zénon et al., 2015), effort-based decision-making (Rudebeck et al., 2006; Rushworth et al.,
2004; Shenhav et al., 2013; Walton et al., 2009, 2004, 2003) and error monitoring (Botvinick
et al., 2004; Dehaene et al., 1994; Gehring and Knight, 2000; Luu and Tucker, 2001; van Veen
et al., 2001; van Veen and Carter, 2002). In that sense, these roles could result from an
interplay with the LC/NA system, as suggested by computational studies (Silvetti et al., 2018).
Still, since most of these results were obtained using long-term modifications of ACC activity
(lesions, transcranial magnetic stimulation) or techniques with low temporal precision
(functional magnetic resonance imaging), the organization and dynamics of the crosstalk
between the LC and the ACC remain hypothetical.
Altogether, our work refines the relation between LC activity and effort production and
its dynamics. Before actions, at the end of the action-triggering process, LC neurons encoded
the subjective difficulty of that process, and possibly the cognitive control allocated. During
sustained actions requiring force exertion, LC neurons encoded the physical effort associated.
Taken together, these modulations suggest a domain-general encoding of effort production,
regardless of the nature of the effort. Considering their timing, these modulations may
promote the mobilization of resources for monitoring the consequences of triggering the
action, both from a motor and executive perspective, such that courses of action requiring
higher effort to be triggered are more likely to succeed. Thus, this works supports a key role
of the LC in effort processing.
145
References Alnæs, D., Sneve, M.H., Espeseth, T., Endestad, T., van de
Pavert, S.H.P., Laeng, B., 2014. Pupil size signals mental
effort deployed during multiple object tracking and predicts
brain activity in the dorsal attention network and the locus
coeruleus. J. Vis. 14. https://doi.org/10.1167/14.4.1
Andrade, L.S., Kanitz, A.C., Häfele, M.S., Schaun, G.Z., Pinto,
S.S., Alberton, C.L., 2020. Relationship between Oxygen
Uptake, Heart Rate, and Perceived Effort in an Aquatic
Incremental Test in Older Women. Int. J. Environ. Res. Public.
Health 17. https://doi.org/10.3390/ijerph17228324
Arnsten, A.F.T., 2006. Stimulants: Therapeutic actions in
ADHD. Neuropsychopharmacol. Off. Publ. Am. Coll.
Neuropsychopharmacol. 31, 2376–2383.
https://doi.org/10.1038/sj.npp.1301164
Aston-Jones, G., 2005. Brain structures and receptors
involved in alertness. Sleep Med. 6, S3–S7.
https://doi.org/10.1016/S1389-9457(05)80002-4
Aston-Jones, G., Ennis, M., Pieribone, V.A., Nickell, W.T.,
Shipley, M.T., 1986. The brain nucleus locus coeruleus:
restricted afferent control of a broad efferent network.
Science 234, 734–737.
https://doi.org/10.1126/science.3775363
Aston-Jones, G., Foote, S.L., Segal, M., 1985. Impulse
conduction properties of noradrenergic locus coeruleus
axons projecting to monkey cerebrocortex. Neuroscience 15,
765–777. https://doi.org/10.1016/0306-4522(85)90077-6
Aston-Jones, G., Shipley, M.T., Chouvet, G., Ennis, M., van
Bockstaele, E., Pieribone, V., Shiekhattar, R., Akaoka, H.,
Drolet, G., Astier, B., Charléty, P., Valentino, R.J., Williams,
J.T., 1991. Afferent regulation of locus coeruleus neurons:
anatomy, physiology and pharmacology, in: Progress in Brain
Research. Elsevier, pp. 47–75.
https://doi.org/10.1016/S0079-6123(08)63799-1
Bari, A., Xu, S., Pignatelli, M., Takeuchi, D., Feng, J., Li, Y.,
Tonegawa, S., 2020. Differential attentional control
mechanisms by two distinct noradrenergic coeruleo-frontal
cortical pathways. Proc. Natl. Acad. Sci. U. S. A. 117, 29080–
29089. https://doi.org/10.1073/pnas.2015635117
Beatty, J., 1982. Task-evoked pupillary responses, processing
load, and the structure of processing resources. Psychol.
Bull. 91, 276–292. https://doi.org/10.1037/0033-
2909.91.2.276
Borderies, N., Bornert, P., Gilardeau, S., Bouret, S., 2020.
Pharmacological evidence for the implication of
noradrenaline in effort. PLOS Biol. 18, e3000793.
https://doi.org/10.1371/journal.pbio.3000793
Bouret, S., Ravel, S., Richmond, B.J., 2012. Complementary
neural correlates of motivation in dopaminergic and
noradrenergic neurons of monkeys. Front. Behav. Neurosci.
6. https://doi.org/10.3389/fnbeh.2012.00040
Bouret, S., Sara, S.J., 2005. Network reset: a simplified
overarching theory of locus coeruleus noradrenaline
function. Trends Neurosci. 28, 574–582.
https://doi.org/10.1016/j.tins.2005.09.002
Bowman, E.M., Aigner, T.G., Richmond, B.J., 1996. Neural
signals in the monkey ventral striatum related to motivation
for juice and cocaine rewards. J. Neurophysiol. 75, 1061–
1073. https://doi.org/10.1152/jn.1996.75.3.1061
Brink, R.L. van den, Pfeffer, T., Warren, C.M., Murphy, P.R.,
Tona, K.-D., Wee, N.J.A. van der, Giltay, E., Noorden, M.S.
van, Rombouts, S.A.R.B., Donner, T.H., Nieuwenhuis, S.,
2016. Catecholaminergic Neuromodulation Shapes Intrinsic
MRI Functional Connectivity in the Human Brain. J. Neurosci.
36, 7865–7876. https://doi.org/10.1523/JNEUROSCI.0744-
16.2016
Buffalari, D.M., Grace, A.A., 2007. Noradrenergic modulation
of basolateral amygdala neuronal activity: opposing
influences of alpha-2 and beta receptor activation. J.
Neurosci. Off. J. Soc. Neurosci. 27, 12358–12366.
https://doi.org/10.1523/JNEUROSCI.2007-07.2007
Carroll, D., Turner, J.R., Hellawell, J.C., 1986. Heart Rate and
Oxygen Consumption during Active Psychological Challenge:
The Effects of Level of Difficulty. Psychophysiology 23, 174–
181. https://doi.org/10.1111/j.1469-8986.1986.tb00613.x
Carstens, E., Leah, J., Lechner, J., Zimmermann, M., 1990.
Demonstration of extensive brainstem projections to medial
and lateral thalamus and hypothalamus in the rat.
Neuroscience 35, 609–626. https://doi.org/10.1016/0306-
4522(90)90333-Y
Carter, M.E., Yizhar, O., Chikahisa, S., Nguyen, H.,
Adamantidis, A., Nishino, S., Deisseroth, K., de Lecea, L.,
2010. Tuning arousal with optogenetic modulation of locus
coeruleus neurons. Nat. Neurosci. 13, 1526–1533.
https://doi.org/10.1038/nn.2682
Cassens, G., Kuruc, A., Roffman, M., Orsulak, P.J.,
Schildkraut, J.J., 1981. Alterations in brain norepinephrine
metabolism and behavior induced by environmental stimuli
previously paired with inescapable shock. Behav. Brain Res.
2, 387–407. https://doi.org/10.1016/0166-4328(81)90020-6
Chamberlain, S.R., Hampshire, A., Müller, U., Rubia, K., Del
Campo, N., Craig, K., Regenthal, R., Suckling, J., Roiser, J.P.,
Grant, J.E., Bullmore, E.T., Robbins, T.W., Sahakian, B.J.,
2009. Atomoxetine modulates right inferior frontal
activation during inhibitory control: a pharmacological
146
functional magnetic resonance imaging study. Biol.
Psychiatry 65, 550–555.
https://doi.org/10.1016/j.biopsych.2008.10.014
Chamberlain, S.R., Müller, U., Blackwell, A.D., Robbins, T.W.,
Sahakian, B.J., 2006. Noradrenergic modulation of working
memory and emotional memory in humans.
Psychopharmacology (Berl.) 188, 397–407.
https://doi.org/10.1007/s00213-006-0391-6
Clark, C.R., Geffen, G.M., Geffen, L.B., 1987. Catecholamines
and attention. I: Animal and clinical studies. Neurosci.
Biobehav. Rev. 11, 341–352. https://doi.org/10.1016/s0149-
7634(87)80006-4
Coradazzi, M., Gulino, R., Fieramosca, F., Falzacappa, L.V.,
Riggi, M., Leanza, G., 2016. Selective noradrenaline
depletion impairs working memory and hippocampal
neurogenesis. Neurobiol. Aging 48, 93–102.
https://doi.org/10.1016/j.neurobiolaging.2016.08.012
Corbetta, M., Patel, G., Shulman, G.L., 2008. The Reorienting
System of the Human Brain: From Environment to Theory of
Mind. Neuron 58, 306–324.
https://doi.org/10.1016/j.neuron.2008.04.017
Coull, J.T., Nobre, A.C., Frith, C.D., 2001. The Noradrenergic
α2 Agonist Clonidine Modulates Behavioural and
Neuroanatomical Correlates of Human Attentional Orienting
and Alerting. Cereb. Cortex 11, 73–84.
https://doi.org/10.1093/cercor/11.1.73
Cousins, M.S., Salamone, J.D., 1994. Nucleus accumbens
dopamine depletions in rats affect relative response
allocation in a novel cost/benefit procedure. Pharmacol.
Biochem. Behav. 49, 85–91. https://doi.org/10.1016/0091-
3057(94)90460-x
Curtis, A.L., Leiser, S.C., Snyder, K., Valentino, R.J., 2012.
Predator stress engages corticotropin-releasing factor and
opioid systems to alter the operating mode of locus
coeruleus norepinephrine neurons. Neuropharmacology 62,
1737–1745.
https://doi.org/10.1016/j.neuropharm.2011.11.020
Da Silva, D.F., Mohammad, S., Hutchinson, K.A., Adamo, K.B.,
2020. Cross-Validation of Ratings of Perceived Exertion
Derived from Heart Rate Target Ranges Recommended for
Pregnant Women. Int. J. Exerc. Sci. 13, 1340–1351.
de Gee, J.W., Knapen, T., Donner, T.H., 2014. Decision-
related pupil dilation reflects upcoming choice and individual
bias. Proc. Natl. Acad. Sci. 111, E618–E625.
https://doi.org/10.1073/pnas.1317557111
Dmochowski, J.P., Norcia, A.M., 2015. Cortical Components
of Reaction-Time during Perceptual Decisions in Humans.
PLOS ONE 10, e0143339.
https://doi.org/10.1371/journal.pone.0143339
Doya, K., 2002. Metalearning and neuromodulation. Neural
Netw. 15, 495–506. https://doi.org/10.1016/S0893-
6080(02)00044-8
Einhäuser, W., Koch, C., Carter, O.L., 2010. Pupil Dilation
Betrays the Timing of Decisions. Front. Hum. Neurosci. 4, 18.
https://doi.org/10.3389/fnhum.2010.00018
Ellis, K.A., Nathan, P.J., 2001. The pharmacology of human
working memory. Int. J. Neuropsychopharmacol. 4, 299–
313. https://doi.org/10.1017/S1461145701002541
Ennis, M., Aston-Jones, G., 1988. Activation of locus
coeruleus from nucleus paragigantocellularis: a new
excitatory amino acid pathway in brain. J. Neurosci. 8, 3644–
3657. https://doi.org/10.1523/JNEUROSCI.08-10-
03644.1988
Ennis, M., Aston-Jones, G., 1986. Evidence for self- and
neighbor-mediated postactivation inhibition of locus
coeruleus neurons. Brain Res. 374, 299–305.
https://doi.org/10.1016/0006-8993(86)90424-5
Epstein, S., 1994. Integration of the cognitive and the
psychodynamic unconscious. Am. Psychol. 49, 709–724.
https://doi.org/10.1037/0003-066X.49.8.709
Faraone, S.V., Biederman, J., Spencer, T., Michelson, D.,
Adler, L., Reimherr, F., Seidman, L., 2005. Atomoxetine and
Stroop Task Performance in Adult Attention-
Deficit/Hyperactivity Disorder. J. Child Adolesc.
Psychopharmacol. 15, 664–670.
https://doi.org/10.1089/cap.2005.15.664
Fernando, A.B.P., Economidou, D., Theobald, D.E., Zou, M.-
F., Newman, A.H., Spoelder, M., Caprioli, D., Moreno, M.,
Hipólito, L., Aspinall, A.T., Robbins, T.W., Dalley, J.W., 2012.
Modulation of high impulsivity and attentional performance
in rats by selective direct and indirect dopaminergic and
noradrenergic receptor agonists. Psychopharmacology
(Berl.) 219, 341–352. https://doi.org/10.1007/s00213-011-
2408-z
Filevich, E., Kühn, S., Haggard, P., 2013. There Is No Free
Won’t: Antecedent Brain Activity Predicts Decisions to
Inhibit. PLOS ONE 8, e53053.
https://doi.org/10.1371/journal.pone.0053053
Filevich, E., Kühn, S., Haggard, P., 2012. Intentional inhibition
in human action: the power of “no.” Neurosci. Biobehav.
Rev. 36, 1107–1118.
https://doi.org/10.1016/j.neubiorev.2012.01.006
Foote, S.L., Bloom, F.E., Aston-Jones, G., 1983. Nucleus locus
ceruleus: new evidence of anatomical and physiological
specificity. Physiol. Rev. 63, 844–914.
https://doi.org/10.1152/physrev.1983.63.3.844
147
Gallagher, K.M., Fadel, P.J., Strømstad, M., Ide, K., Smith,
S.A., Querry, R.G., Raven, P.B., Secher, N.H., 2001. Effects of
partial neuromuscular blockade on carotid baroreflex
function during exercise in humans. J. Physiol. 533, 861–870.
https://doi.org/10.1111/j.1469-7793.2001.t01-1-00861.x
Gamo, N.J., Wang, M., Arnsten, A.F.T., 2010.
Methylphenidate and Atomoxetine Enhance Prefrontal
Function Through α2-Adrenergic and Dopamine D1
Receptors. J. Am. Acad. Child Adolesc. Psychiatry 49, 1011–
1023. https://doi.org/10.1016/j.jaac.2010.06.015
Gilden, D.L., 1997. Fluctuations in the Time Required for
Elementary Decisions. Psychol. Sci. 8, 296–301.
https://doi.org/10.1111/j.1467-9280.1997.tb00441.x
Gilzenrat, M.S., Nieuwenhuis, S., Jepma, M., Cohen, J.D.,
2010. Pupil diameter tracks changes in control state
predicted by the adaptive gain theory of locus coeruleus
function. Cogn. Affect. Behav. Neurosci. 10, 252–269.
https://doi.org/10.3758/CABN.10.2.252
Griffiths, T.L., Lieder, F., Goodman, N.D., 2015. Rational use
of cognitive resources: levels of analysis between the
computational and the algorithmic. Top. Cogn. Sci. 7, 217–
229. https://doi.org/10.1111/tops.12142
Guedj, C., Monfardini, E., Reynaud, A.J., Farnè, A., Meunier,
M., Hadj-Bouziane, F., 2017. Boosting Norepinephrine
Transmission Triggers Flexible Reconfiguration of Brain
Networks at Rest. Cereb. Cortex N. Y. N 1991 27, 4691–4700.
https://doi.org/10.1093/cercor/bhw262
Hernaus, D., Casales Santa, M.M., Offermann, J.S., Van
Amelsvoort, T., 2017. Noradrenaline transporter blockade
increases fronto-parietal functional connectivity relevant for
working memory. Eur. Neuropsychopharmacol. 27, 399–
410. https://doi.org/10.1016/j.euroneuro.2017.02.004
Hess, E.H., Polt, J.M., 1964. Pupil Size in Relation to Mental
Activity during Simple Problem-Solving. Science 143, 1190–
1192. https://doi.org/10.1126/science.143.3611.1190
Hori, Y., Minamimoto, T., Kimura, M., 2009. Neuronal
Encoding of Reward Value and Direction of Actions in the
Primate Putamen. J. Neurophysiol. 102, 3530–3543.
https://doi.org/10.1152/jn.00104.2009
Hull, C.L., 1943. Principles of behavior: an introduction to
behavior theory, Principles of behavior: an introduction to
behavior theory. Appleton-Century, Oxford, England.
Hunt, L.T., Kolling, N., Soltani, A., Woolrich, M.W.,
Rushworth, M.F.S., Behrens, T.E.J., 2012. Mechanisms
underlying cortical activity during value-guided choice. Nat.
Neurosci. 15, 470–476. https://doi.org/10.1038/nn.3017
Jahn, C.I., Gilardeau, S., Varazzani, C., Blain, B., Sallet, J.,
Walton, M.E., Bouret, S., 2018. Dual contributions of
noradrenaline to behavioural flexibility and motivation.
Psychopharmacology (Berl.) 235, 2687–2702.
https://doi.org/10.1007/s00213-018-4963-z
Jahn, C.I., Varazzani, C., Sallet, J., Walton, M.E., Bouret, S.,
2020. Noradrenergic But Not Dopaminergic Neurons Signal
Task State Changes and Predict Reengagement After a
Failure. Cereb. Cortex N. Y. N 1991 30, 4979–4994.
https://doi.org/10.1093/cercor/bhaa089
Jones, B.E., Moore, R.Y., 1977. Ascending projections of the
locus coeruleus in the rat. II. Autoradiographic study. Brain
Res. 127, 23–53. https://doi.org/10.1016/0006-
8993(77)90378-X
Joshi, S., Li, Y., Kalwani, R.M., Gold, J.I., 2016. Relationships
between Pupil Diameter and Neuronal Activity in the Locus
Coeruleus, Colliculi, and Cingulate Cortex. Neuron 89, 221–
234. https://doi.org/10.1016/j.neuron.2015.11.028
Kahneman, D., 1973. Attention and effort. Prentice-Hall,
Englewood Cliffs, N.J.
Kahneman, D., Beatty, J., 1966. Pupil diameter and load on
memory. Science 154, 1583–1585.
https://doi.org/10.1126/science.154.3756.1583
Kalwani, R.M., Joshi, S., Gold, J.I., 2014. Phasic Activation of
Individual Neurons in the Locus Ceruleus/Subceruleus
Complex of Monkeys Reflects Rewarded Decisions to Go But
Not Stop. J. Neurosci. 34, 13656–13669.
https://doi.org/10.1523/JNEUROSCI.2566-14.2014
Koob, G.F., 1999. Corticotropin-releasing factor,
norepinephrine, and stress. Biol. Psychiatry 46, 1167–1180.
https://doi.org/10.1016/s0006-3223(99)00164-x
Kool, W., Botvinick, M., 2018. Mental labour. Nat. Hum.
Behav. 2, 899–908. https://doi.org/10.1038/s41562-018-
0401-9
Kool, W., Botvinick, M., 2014. A labor/leisure tradeoff in
cognitive control. J. Exp. Psychol. Gen. 143, 131–141.
https://doi.org/10.1037/a0031048
Kool, W., McGuire, J.T., Rosen, Z.B., Botvinick, M.M., 2010.
Decision making and the avoidance of cognitive demand. J.
Exp. Psychol. Gen. 139, 665–682.
https://doi.org/10.1037/a0020198
Krajbich, I., Armel, C., Rangel, A., 2010. Visual fixations and
the computation and comparison of value in simple choice.
Nat. Neurosci. 13, 1292–1298.
https://doi.org/10.1038/nn.2635
148
Krajbich, I., Rangel, A., 2011. Multialternative drift-diffusion
model predicts the relationship between visual fixations and
choice in value-based decisions. Proc. Natl. Acad. Sci. 108,
13852–13857. https://doi.org/10.1073/pnas.1101328108
La Camera, G., Bouret, S., Richmond, B.J., 2018.
Contributions of Lateral and Orbital Frontal Regions to
Abstract Rule Acquisition and Reversal in Monkeys. Front.
Neurosci. 12, 165.
https://doi.org/10.3389/fnins.2018.00165
Lapiz, M.D.S., Bondi, C.O., Morilak, D.A., 2007. Chronic
treatment with desipramine improves cognitive
performance of rats in an attentional set-shifting test.
Neuropsychopharmacol. Off. Publ. Am. Coll.
Neuropsychopharmacol. 32, 1000–1010.
https://doi.org/10.1038/sj.npp.1301235
Lapiz, M.D.S., Morilak, D.A., 2006. Noradrenergic
modulation of cognitive function in rat medial prefrontal
cortex as measured by attentional set shifting capability.
Neuroscience 137, 1039–1049.
https://doi.org/10.1016/j.neuroscience.2005.09.031
Le Bouc, R., Rigoux, L., Schmidt, L., Degos, B., Welter, M.-L.,
Vidailhet, M., Daunizeau, J., Pessiglione, M., 2016.
Computational Dissection of Dopamine Motor and
Motivational Functions in Humans. J. Neurosci. 36, 6623–
6633. https://doi.org/10.1523/JNEUROSCI.3078-15.2016
Lempert, K.M., Chen, Y.L., Fleming, S.M., 2015. Relating Pupil
Dilation and Metacognitive Confidence during Auditory
Decision-Making. PloS One 10, e0126588.
https://doi.org/10.1371/journal.pone.0126588
Li, B.-M., Mao, Z.-M., Wang, M., Mei, Z.-T., 1999. Alpha-2
Adrenergic Modulation of Prefrontal Cortical Neuronal
Activity Related to Spatial Working Memory in Monkeys.
Neuropsychopharmacology 21, 601–610.
https://doi.org/10.1016/S0893-133X(99)00070-6
Lieder, F., Griffiths, T.L., 2017. Strategy selection as rational
metareasoning. Psychol. Rev. 124, 762–794.
https://doi.org/10.1037/rev0000075
Lieder, F., Griffiths, T.L., Goodman, N.D., 2012. Burn-in, bias,
and the rationality of anchoring. Presented at the NIPS.
Lieder, F., Plunkett, D., Hamrick, J.B., Russell, S.J., Hay, N.J.,
Griffiths, T.L., 2014. Algorithm selection by rational
metareasoning as a model of human strategy selection, in:
Proceedings of the 27th International Conference on Neural
Information Processing Systems - Volume 2, NIPS’14. MIT
Press, Cambridge, MA, USA, pp. 2870–2878.
Mao, Z.-M., Arnsten, A.F.T., Li, B.-M., 1999. Local infusion of
an α-1 adrenergic agonist into the prefrontal cortex impairs
spatial working memory performance in monkeys. Biol.
Psychiatry 46, 1259–1265. https://doi.org/10.1016/S0006-
3223(99)00139-0
Martínez-Horta, S., Riba, J., de Bobadilla, R.F.,
Pagonabarraga, J., Pascual-Sedano, B., Antonijoan, R.M.,
Romero, S., Mañanas, M.À., García-Sanchez, C., Kulisevsky,
J., 2014. Apathy in Parkinson’s Disease: Neurophysiological
Evidence of Impaired Incentive Processing. J. Neurosci. 34,
5918–5926.
Matsumoto, N., Minamimoto, T., Graybiel, A.M., Kimura, M.,
2001. Neurons in the thalamic CM-Pf complex supply striatal
neurons with information about behaviorally significant
sensory events. J. Neurophysiol. 85, 960–976.
https://doi.org/10.1152/jn.2001.85.2.960
McBride, R.L., Sutin, J., 1976. Projections of the locus
coeruleus and adjacent pontine tegmentum in the cat. J.
Comp. Neurol. 165, 265–284.
https://doi.org/10.1002/cne.901650302
McGAUGHY, J., ROSS, R.S., EICHENBAUM, H., 2008.
NORADRENERGIC, BUT NOT CHOLINERGIC,
DEAFFERENTATION OF PREFRONTAL CORTEX IMPAIRS
ATTENTIONAL SET-SHIFTING. Neuroscience 153, 63–71.
https://doi.org/10.1016/j.neuroscience.2008.01.064
Milosavljevic, M., Malmaud, J., Huth, A., Koch, C., Rangel, A.,
2010. The Drift Diffusion Model Can Account for the
Accuracy and Reaction Time of Value-Based Choices Under
High and Low Time Pressure. SSRN Electron. J.
https://doi.org/10.2139/ssrn.1901533
Minamimoto, T., Hori, Y., Kimura, M., 2009a. Roles of the
thalamic CM–PF complex—Basal ganglia circuit in externally
driven rebias of action. Brain Res. Bull. 78, 75–79.
https://doi.org/10.1016/j.brainresbull.2008.08.013
Minamimoto, T., Hori, Y., Kimura, M., 2005. Complementary
process to response bias in the centromedian nucleus of the
thalamus. Science 308, 1798–1801.
https://doi.org/10.1126/science.1109154
Minamimoto, T., Hori, Y., Richmond, B.J., 2012. Is Working
More Costly than Waiting in Monkeys? PLOS ONE 7, e48434.
https://doi.org/10.1371/journal.pone.0048434
Minamimoto, T., Kimura, M., 2002. Participation of the
thalamic CM-Pf complex in attentional orienting. J.
Neurophysiol. 87, 3090–3101.
https://doi.org/10.1152/jn.2002.87.6.3090
Minamimoto, T., La Camera, G., Richmond, B.J., 2009b.
Measuring and modeling the interaction among reward size,
delay to reward, and satiation level on motivation in
monkeys. J. Neurophysiol. 101, 437–447.
https://doi.org/10.1152/jn.90959.2008
149
Morris, R., Bornstein, A., Shenhav, A. (Eds.), 2019. Goal-
directed decision making: computations and neural circuits.
Academic Press, Elsevier, London San Diego, CA Cambridge.
Murphy, P.R., Moort, M.L. van, Nieuwenhuis, S., 2016. The
Pupillary Orienting Response Predicts Adaptive Behavioral
Adjustment after Errors. PLOS ONE 11, e0151763.
https://doi.org/10.1371/journal.pone.0151763
Murphy, P.R., O’Connell, R.G., O’Sullivan, M., Robertson, I.H.,
Balsters, J.H., 2014. Pupil diameter covaries with BOLD
activity in human locus coeruleus: Pupil Diameter and Locus
Coeruleus Activity. Hum. Brain Mapp. 35, 4140–4154.
https://doi.org/10.1002/hbm.22466
Nassar, M.R., Rumsey, K.M., Wilson, R.C., Parikh, K., Heasly,
B., Gold, J.I., 2012. Rational regulation of learning dynamics
by pupil-linked arousal systems. Nat. Neurosci. 15, 1040–
1046. https://doi.org/10.1038/nn.3130
Navarra, R., Graf, R., Huang, Y., Logue, S., Comery, T.,
Hughes, Z., Day, M., 2008. Effects of atomoxetine and
methylphenidate on attention and impulsivity in the 5-
choice serial reaction time test. Prog.
Neuropsychopharmacol. Biol. Psychiatry 32, 34–41.
https://doi.org/10.1016/j.pnpbp.2007.06.017
Navon, D., 1984. Resources--a theoretical soup stone?
Psychol. Rev. 91, 216–234. https://doi.org/10.1037/0033-
295X.91.2.216
Nieuwenhuis, S., Aston-Jones, G., Cohen, J.D., 2005. Decision
making, the P3, and the locus coeruleus-norepinephrine
system. Psychol. Bull. 131, 510–532.
https://doi.org/10.1037/0033-2909.131.4.510
Pessiglione, M., Vinckier, F., Bouret, S., Daunizeau, J., Le
Bouc, R., 2018. Why not try harder? Computational approach
to motivation deficits in neuro-psychiatric diseases. Brain
141, 629–650. https://doi.org/10.1093/brain/awx278
Philiastides, M.G., Ratcliff, R., 2013. Influence of Branding on
Preference-Based Decision Making. Psychol. Sci. 24, 1208–
1215. https://doi.org/10.1177/0956797612470701
Phillips, P.E.M., Walton, M.E., Jhou, T.C., 2007. Calculating
utility: preclinical evidence for cost-benefit analysis by
mesolimbic dopamine. Psychopharmacology (Berl.) 191,
483–495. https://doi.org/10.1007/s00213-006-0626-6
Reynaud, A.J., Froesel, M., Guedj, C., Ben Hadj Hassen, S.,
Cléry, J., Meunier, M., Ben Hamed, S., Hadj-Bouziane, F.,
2019. Atomoxetine improves attentional orienting in a
predictive context. Neuropharmacology 150, 59–69.
https://doi.org/10.1016/j.neuropharm.2019.03.012
Robinson, E.S.J., 2012. Blockade of noradrenaline re-uptake
sites improves accuracy and impulse control in rats
performing a five-choice serial reaction time tasks.
Psychopharmacology (Berl.) 219, 303–312.
https://doi.org/10.1007/s00213-011-2420-3
Robinson, E.S.J., Eagle, D.M., Mar, A.C., Bari, A., Banerjee, G.,
Jiang, X., Dalley, J.W., Robbins, T.W., 2008. Similar effects of
the selective noradrenaline reuptake inhibitor atomoxetine
on three distinct forms of impulsivity in the rat.
Neuropsychopharmacol. Off. Publ. Am. Coll.
Neuropsychopharmacol. 33, 1028–1037.
https://doi.org/10.1038/sj.npp.1301487
Rudebeck, P.H., Walton, M.E., Smyth, A.N., Bannerman,
D.M., Rushworth, M.F.S., 2006. Separate neural pathways
process different decision costs. Nat. Neurosci. 9, 1161–
1168. https://doi.org/10.1038/nn1756
Rushworth, M.F.S., Walton, M.E., Kennerley, S.W.,
Bannerman, D.M., 2004. Action sets and decisions in the
medial frontal cortex. Trends Cogn. Sci. 8, 410–417.
https://doi.org/10.1016/j.tics.2004.07.009
Schwarz, L.A., Miyamichi, K., Gao, X.J., Beier, K.T.,
Weissbourd, B., DeLoach, K.E., Ren, J., Ibanes, S., Malenka,
R.C., Kremer, E.J., Luo, L., 2015. Viral-genetic tracing of the
input-output organization of a central noradrenaline circuit.
Nature 524, 88–92. https://doi.org/10.1038/nature14600
Sellaro, R., van Leusden, J.W.R., Tona, K.-D., Verkuil, B.,
Nieuwenhuis, S., Colzato, L.S., 2015. Transcutaneous Vagus
Nerve Stimulation Enhances Post-error Slowing. J. Cogn.
Neurosci. 27, 2126–2132.
https://doi.org/10.1162/jocn_a_00851
Shea-Brown, E., Gilzenrat, M.S., Cohen, J.D., 2008.
Optimization of decision making in multilayer networks: the
role of locus coeruleus. Neural Comput. 20, 2863–2894.
https://doi.org/10.1162/neco.2008.03-07-487
Shenhav, A., Botvinick, M.M., Cohen, J.D., 2013. The
Expected Value of Control: An Integrative Theory of Anterior
Cingulate Cortex Function. Neuron 79, 217–240.
https://doi.org/10.1016/j.neuron.2013.07.007
Shenhav, A., Musslick, S., Lieder, F., Kool, W., Griffiths, T.L.,
Cohen, J.D., Botvinick, M.M., 2017. Toward a Rational and
Mechanistic Account of Mental Effort. Annu. Rev. Neurosci.
40, 99–124. https://doi.org/10.1146/annurev-neuro-
072116-031526
Shiffrin, R.M., Schneider, W., 1977. Controlled and automatic
human information processing: II. Perceptual learning,
automatic attending and a general theory. Psychol. Rev. 84,
127–190. https://doi.org/10.1037/0033-295X.84.2.127
Silvetti, M., Vassena, E., Abrahamse, E., Verguts, T., 2018.
Dorsal anterior cingulate-brainstem ensemble as a
reinforcement meta-learner. PLOS Comput. Biol. 14,
e1006370. https://doi.org/10.1371/journal.pcbi.1006370
150
Simpson, H.M., Hale, S.M., 1969. Pupillary changes during a
decision-making task. Percept. Mot. Skills 29, 495–498.
https://doi.org/10.2466/pms.1969.29.2.495
Sims, J., Carroll, D., 1990. Cardiovascular and Metabolic
Activity at Rest and During Psychological and Physical
Challenge in Normotensives and Subjects With Mildly
Elevated Blood Pressure. Psychophysiology 27, 149–156.
https://doi.org/10.1111/j.1469-8986.1990.tb00366.x
Sims, J., Carroll, D., Turner, J.R., Hewitt, J.K., 1988. Cardiac
and Metabolic Activity in Mild Hypertensive and
Normotensive Subjects. Psychophysiology 25, 172–178.
https://doi.org/10.1111/j.1469-8986.1988.tb00982.x
Sloman, S.A., 1996. The empirical case for two systems of
reasoning. Psychol. Bull. 119, 3–22.
https://doi.org/10.1037/0033-2909.119.1.3
Stevens, J.R., Rosati, A.G., Ross, K.R., Hauser, M.D., 2005. Will
travel for food: spatial discounting in two new world
monkeys. Curr. Biol. CB 15, 1855–1860.
https://doi.org/10.1016/j.cub.2005.09.016
Sved, A.F., Cano, G., Passerin, A.M., Rabin, B.S., 2002. The
locus coeruleus, Barrington’s nucleus, and neural circuits of
stress. Physiol. Behav. 77, 737–742.
https://doi.org/10.1016/s0031-9384(02)00927-7
Tait, D.S., Brown, V.J., Farovik, A., Theobald, D.E., Dalley,
J.W., Robbins, T.W., 2007. Lesions of the dorsal
noradrenergic bundle impair attentional set-shifting in the
rat. Eur. J. Neurosci. 25, 3719–3724.
https://doi.org/10.1111/j.1460-9568.2007.05612.x
Tibana, R., Manuel Frade de Sousa, N., Prestes, J., da Cunha
Nascimento, D., Ernesto, C., Falk Neto, J., Kennedy, M.,
Azevedo Voltarelli, F., 2019. Is Perceived Exertion a Useful
Indicator of the Metabolic and Cardiovascular Responses to
a Metabolic Conditioning Session of Functional Fitness?
Sports 7, 161. https://doi.org/10.3390/sports7070161
Tsunematsu, S., 2001. Effort- and time-cost effects on
demand curves for food by pigeons under short session
closed economies. Behav. Processes 53, 47–56.
https://doi.org/10.1016/S0376-6357(00)00147-9
Turner, J.R., Carroll, D., 1985. Heart Rate and Oxygen
Consumption during Mental Arithmetic, a Video Game, and
Graded Exercise: Further Evidence of Metabolically-
Exaggerated Cardiac Adjustments? Psychophysiology 22,
261–267. https://doi.org/10.1111/j.1469-
8986.1985.tb01597.x
Turner, J.R., Carroll, D., Courtney, H., 1983. Cardiac and
Metabolic Responses to “Space Invaders”: An Instance of
Metabolically-Exaggerated Cardiac Adjustment?
Psychophysiology 20, 544–549.
https://doi.org/10.1111/j.1469-8986.1983.tb03010.x
Tzavara, E.T., Bymaster, F.P., Overshiner, C.D., Davis, R.J.,
Perry, K.W., Wolff, M., McKinzie, D.L., Witkin, J.M., Nomikos,
G.G., 2006. Procholinergic and memory enhancing
properties of the selective norepinephrine uptake inhibitor
atomoxetine. Mol. Psychiatry 11, 187–195.
https://doi.org/10.1038/sj.mp.4001763
Valentino, R.J., Page, M.E., Curtis, A.L., 1991. Activation of
noradrenergic locus coeruleus neurons by hemodynamic
stress is due to local release of corticotropin-releasing factor.
Brain Res. 555, 25–34. https://doi.org/10.1016/0006-
8993(91)90855-p
Van Bockstaele, E.J., Aston-Jones, G., 1995. Integration in the
Ventral Medulla and Coordination of Sympathetic, Pain and
Arousal Functions. Clin. Exp. Hypertens. 17, 153–165.
https://doi.org/10.3109/10641969509087062
Van Bockstaele, E.J., Bajic, D., Proudfit, H., Valentino, R.J.,
2001. Topographic architecture of stress-related pathways
targeting the noradrenergic locus coeruleus. Physiol. Behav.
73, 273–283. https://doi.org/10.1016/s0031-
9384(01)00448-6
Varazzani, C., San-Galli, A., Gilardeau, S., Bouret, S., 2015.
Noradrenaline and Dopamine Neurons in the Reward/Effort
Trade-Off: A Direct Electrophysiological Comparison in
Behaving Monkeys. J. Neurosci. 35, 7866–7877.
https://doi.org/10.1523/JNEUROSCI.0454-15.2015
Wahn, B., Ferris, D.P., Hairston, W.D., König, P., 2016. Pupil
Sizes Scale with Attentional Load and Task Experience in a
Multiple Object Tracking Task. PloS One 11, e0168087.
https://doi.org/10.1371/journal.pone.0168087
Walton, M.E., Bannerman, D.M., Alterescu, K., Rushworth,
M.F.S., 2003. Functional Specialization within Medial Frontal
Cortex of the Anterior Cingulate for Evaluating Effort-Related
Decisions. J. Neurosci. 23, 6475–6479.
https://doi.org/10.1523/JNEUROSCI.23-16-06475.2003
Walton, M.E., Devlin, J.T., Rushworth, M.F.S., 2004.
Interactions between decision making and performance
monitoring within prefrontal cortex. Nat. Neurosci. 7, 1259–
1265. https://doi.org/10.1038/nn1339
Walton, M.E., Groves, J., Jennings, K.A., Croxson, P.L., Sharp,
T., Rushworth, M.F.S., Bannerman, D.M., 2009. Comparing
the role of the anterior cingulate cortex and 6-
hydroxydopamine nucleus accumbens lesions on operant
effort-based decision making. Eur. J. Neurosci. 29, 1678–
1691. https://doi.org/10.1111/j.1460-9568.2009.06726.x
Walton, M.E., Kennerley, S.W., Bannerman, D.M., Phillips,
P.E.M., Rushworth, M.F.S., 2006. Weighing up the benefits
of work: behavioral and neural analyses of effort-related
decision making. Neural Netw. Off. J. Int. Neural Netw. Soc.
19, 1302–1314.
https://doi.org/10.1016/j.neunet.2006.03.005
151
Wang, M., Ramos, B.P., Paspalas, C.D., Shu, Y., Simen, A.,
Duque, A., Vijayraghavan, S., Brennan, A., Dudley, A., Nou,
E., Mazer, J.A., McCormick, D.A., Arnsten, A.F.T., 2007. α2A-
Adrenoceptors Strengthen Working Memory Networks by
Inhibiting cAMP-HCN Channel Signaling in Prefrontal Cortex.
Cell 129, 397–410.
https://doi.org/10.1016/j.cell.2007.03.015
White, O., Davare, M., Andres, M., Olivier, E., 2013. The role
of left supplementary motor area in grip force scaling. PloS
One 8, e83812.
https://doi.org/10.1371/journal.pone.0083812
Widge, A.S., Zorowitz, S., Basu, I., Paulk, A.C., Cash, S.S.,
Eskandar, E.N., Deckersbach, T., Miller, E.K., Dougherty, D.D.,
2019. Deep brain stimulation of the internal capsule
enhances human cognitive control and prefrontal cortex
function. Nat. Commun. 10, 1536.
https://doi.org/10.1038/s41467-019-09557-4
Witte, E.A., Marrocco, R.T., 1997. Alteration of brain
noradrenergic activity in rhesus monkeys affects the alerting
component of covert orienting. Psychopharmacology (Berl.)
132, 315–323. https://doi.org/10.1007/s002130050351
Wright, T.J., Boot, W.R., Morgan, C.S., 2013. Pupillary
response predicts multiple object tracking load, error rate,
and conscientiousness, but not inattentional blindness. Acta
Psychol. (Amst.) 144, 6–11.
https://doi.org/10.1016/j.actpsy.2013.04.018
Zénon, A., Sidibé, M., Olivier, E., 2015. Disrupting the
Supplementary Motor Area Makes Physical Effort Appear
Less Effortful. J. Neurosci. 35, 8737–8744.
https://doi.org/10.1523/JNEUROSCI.3789-14.2015
Zénon, A., Sidibé, M., Olivier, E., 2014. Pupil size variations
correlate with physical effort perception. Front. Behav.
Neurosci. 8, 286.
https://doi.org/10.3389/fnbeh.2014.00286
Zipf, G.K., 1949. Human behavior and the principle of least
effort, Human behavior and the principle of least effort.
Addison-Wesley Press, Oxford, England.
Experimental results
152
Supplementary material
Figure S1: Force exertion in the Force-discounting task across time after action onset by (A)
Monkey D and (B) Monkey A, by force categories. The thin lines represent the mean and SEM
of the exerted force at each time point.
Experimental results
153
2. Additional results
The study presented above focused on the activity of LC neurons around action onset,
arguing for a role of the LC/NA system in reporting on the amount of cognitive control that
was invested in overcoming the cost of triggering a costly action. In the following part, I report
a few additional findings regarding the Delay-Discounting task, considering that this dataset
was never the subject of any other publication. In particular, a first subpart will focus on the
activity of LC neurons around cue onset and the outcome. Then, a second subpart will focus
on the temporal dynamics of the responses to the three task events, and of the modulation of
the rate of LC neurons by task parameters and behavioral variables. Third, a last subpart will
detail findings regarding the heterogeneity of LC neurons in the magnitude of their responses
to task events, and in the strength of the encoding of task parameters and behavioral
parameters. Finally, I will briefly discuss such findings, although more extensive discussion will
be provided in the very last part of this manuscript.
a) Activity of LC neurons around cue onset and the
outcome
(1) LC neurons responded to cue onset and the outcome
Besides displaying a strong response to action onset, LC neurons displayed a phasic
response to cue onset and to the outcome (reward delivery). I quantified the changes in firing
of each neuron by comparing spike counts in a fixed window around these two events of
interest ([0;300ms] after cue onset and [0;300ms] after outcome) and baseline activity ([0 ;
300ms] before cue onset). Practically, for each neuron, I performed a paired t-test between
the spike count computed in the window of interest and the spike count in the baseline
window. I then evaluated the reliability of rate changes at the population level using a second-
level analysis (t-test on the T values from neuron-by neuron t-tests on rate). As shown on
Figure 41, the population of LC neurons displayed a significant activation in response to cue
onset and the outcome (cue: t(74)=5.37; p<0.001; outcome: t(74)=6.94; p<0.001). Out of 75
LC units, 23 (30.6%) displayed a significant response to the cue, and 39 (52%) at the trial
outcome. On average, the LC neurons increased their activity (compared to pre-cue baseline
firing) by 21.7% after cue onset and by 25.4% after the outcome.
Experimental results
154
Figure 41: Summary of the mean variations in rate for the 75 recorded neurons around cue
onset (A, B) and the outcome (C, D).
(A,C): Sliding window analyses of rate. Each line of the plot corresponds to the mean rate of
one individual neuron (y-axis: neuron number, N°) across trials at each time point (middle of
the time window used for rate assessment). The color code used is displayed on the right.
(A) Sliding window analysis of rate performed around cue onset. Just after cue onset, rate
increased for most neurons.
(B) Density plot of the T values from the T tests comparing the neuron by neuron rate in the
[0;300ms] epoch after cue onset to baseline rate in the [-300;0] epoch before cue onset. In that
Experimental results
155
epoch, T values are significantly greater than zero, indicating an increase in rate across the
population compared to baseline rate.
(C) Sliding window analysis of rate performed around the outcome (reward delivery). Just after
the outcome, rate increases for most neurons compared to the baseline rate.
(D) Density plot of the T values from the T tests comparing the neuron by neuron rate in the
[0;300ms] epoch after the outcome to baseline rate in the [-300;0] epoch before cue onset. In
the early epoch, T values are significantly greater than zero, indicating an increase in rate
across the population compared to baseline rate.
(2) Modulation of LC responses by task parameters at cue
onset and the outcome
Besides the strong changes in activity related to cue onset and the outcome, the firing
of LC neurons around cue onset and the outcome was also modulated across the 9 task
conditions, defined both by the reward magnitude and the delay duration (Figure 42).
I quantified the parametric effects of Reward and Delay (three levels each) on spike
counts around these events of interest using generalized linear models (GLMs). I counted
spikes in a [0;700ms] window after cue onset and a [0;500ms] window after trial outcome,
based on raster inspection. Spike counts of each unit were fitted with a GLM to evaluate the
relative weight of Reward (βR) and Delay (βD) modulation. For each event, I used a similar
approach as that described above (Bornert and Bouret, in prep) for event-related changes in
firing rate: a neuron-by-neuron analysis to evaluate the strength of the parametric effects in
individual units and a second-level population analysis to capture the coherence of these
effects across the population (Figure 42).
At cue onset, LC neurons displayed a significant modulation by task parameters (Figure
42A). The positive influence of reward was significant for 45/75 neurons. It was also significant
at the population level, since the distribution of beta weights for the Reward parameter (βR)
was significantly positive (t(74)=11.55; p<0.001). Delay had a significant negative effect (βD)
for 33/75 neurons and the second-level analysis confirmed that this effect was significant at
the level of the population (t(74)=-6.57; p<0.001). I used a paired t test to compare the
magnitude of the effects of Reward and Delay (absolute value of βR and βD) and the difference
did not reach significant (p>0.05).
Experimental results
156
Figure 42: Summary of the modulation of rate by task parameters in large time intervals
around cue onset and the outcome
(A) Density plot of the beta values from the GLM for spike count with Reward (beta values in
green) and Delay (beta value in blue) as parameters, in the [0;700ms] epoch after cue onset.
The mean of the beta value for Reward was significantly greater than zero while the mean of
the beta values for Delay was significantly lower than zero i.e. Reward had a positive effect on
spike count and Delay had a negative effect on spike count after cue onset.
(B) Density plot of the beta values from the GLM for spike count with Reward size (beta values
in green) and Delay (beta value in blue) as parameters, in the [0;500ms] epoch after the
outcome. The mean of the beta values for Reward was not significantly different from zero
while the mean of the beta values for Delay was significantly greater than zero i.e. Delay had
a positive effect on spike count after the outcome but we could not find any consistent effect
of Reward.
At the outcome, the population of LC neurons showed only a significant positive
modulation by Delay (t(74)=6.54; p<0.001) but the effect of Reward was not significant
(p>0.05) (Figure 42B). At the level of individual neurons, only 8/75 neurons showed a
significant effect of Reward, and it was negative for 3 of them. By contrast, 41/75 neurons
showed a significant positive effect of Delay. On average, the magnitude of the Delay
modulation was greater than that of Reward modulation (t(74)=8.45; p<0.001).
In summary, the firing of LC neurons around cue onset and the outcome displayed a
significant modulation by task parameters. At the cue, firing rate was modulated positively by
(B)
Experimental results
157
the Reward and negatively by the Delay, with no significant difference between the effects’
magnitude. At the trial outcome, the influence of task parameters on LC activity reversed:
Reward decreased LC firing whereas Delay increased it. Moreover, the relative magnitude of
the Delay modulation increased compared to that of Reward. In that, the modulation of LC
activity around the outcome of the trial resembles that observed just prior to action onset
(Bornert and Bouret, in prep).
(3) Modulation of LC activity by behavior
(a) Cue-evoked activity and willingness to work
At the time of cue onset, LC neurons encoded information about Reward: its size,
positively, and its Delay, negatively. Given the similarity between the influence of task
parameters on LC activity and willingness to work, I tested the direct relation between these
two variables. Note that willingness to work here refers to the proportion of acceptance of
each condition, by session. Indeed, I used acceptance rates, a measure of the reluctance of
the monkeys to engage in the trials, as a proxy for the subjective cost of engaging in specific
conditions, and tested whether LC neurons encoded this measure at all, and over and above
task parameters (Reward and Delay).
First, I regressed the spike count in a large post-cue epoch ([0;700ms]) on willingness
to work alone, using a simple GLM for each individual neuron. I had to exclude 12/75 neurons
recorded during sessions in which the monkeys had reached 100% engagement. Nine of the
remaining 54 neurons exhibited a significant positive effect of willingness to work, and a
second level analysis revealed that this positive effect was consistent across the population (t
test: t(53)=8.756; p<0.001). I then tested whether this effect was merely confounded by a joint
relation with task parameters, by including Reward and Delay as covariates of willingness to
work in the regression. The positive effect of willingness to work was significant for 12/54
units and the second level analysis indicated that the positive effect was still marginally
significant across the population (t test: t(53)=1.943; p=0.057).
Importantly, these effects were not present in the baseline rate i.e. before cue onset,
meaning that the encoding of willingness to work was dependent on the information about
the condition of the trial provided by the cue. Given the marginal result of the statistical test,
it remains difficult to draw a strong conclusion, yet such results are compatible with the idea
that the activity of LC neurons at the cue is related to the monkeys’ willingness to work.
Experimental results
158
(b) Real delay and outcome activity
In our task, Delay condition (1,2 or 3), an information provided by the cue, was only
partially informative of the time that would elapse between action onset and reward delivery.
Indeed, each Delay category corresponded to a time range in which the actual delay could fall.
To assess the influence of the amount of time elapsed between action and reward delivery
(actual delay) on the magnitude of the outcome (reward)-evoked activity, I regressed outcome
evoked firing rates in the [0;500ms] epoch after the outcome using a GLM with the actual
delay (time elapsed). I included Delay category (1, 2 or 3) as co-regressor, to examine the
influence of influence of actual delay (experienced at the outcome) beyond the Delay category
(known from trial onset). In these conditions, 9/75 neurons exhibited a significant effect of
actual delay (5 negative, 4 positive), and the second level analysis indicated that actual delay
did not have a consistent effect at the population level (t test against zero: p>0.05).
b) Dissection of the temporal dynamics of parameter
encoding relative to activity changes
I used a sliding window procedure to capture the dynamics of firing rate changes not
only within a trial (how firing rate changed from baseline in response to individual task event)
but also across trials (how event-related firing rate varied across the 9 task conditions, defined
by levels of Reward and Delay).
Practically, for rate change latencies, I used the rate computed in the sliding window
procedure (see methods section from Bornert and Bouret, in prep) and I performed a
systematic one-sided Wilcoxon test (alternative hypothesis: greater to detect the increase in
rate) between the rate in each window and the baseline rate (10 first windows of the sliding
window procedure). I determined the position of the first three consecutive windows with a
significant result. The latencies reported correspond to the middle of the first of these three
windows. For rate peak latency, I looked for the position of the window of the rate assessment
sliding window procedure with the highest spike count. I report the timing of the middle of
that window. Regarding the latency to the detection of a modulation of rate by task
parameters and behavioral variables, I used the beta value from the GLM sliding window
procedure (see methods section from Bornert and Bouret, in prep). I identified the latency to
Experimental results
159
the effects of each parameter by determining the position of the first three consecutive
windows of time with a significant coefficient for this parameter. I then report the middle of
the first of these windows as latency to the effect. The results of these analyses, i.e. latencies,
are provided in Table 1.
Table 1: Latencies of the firing rate changes and task parameter encoding at cue onset, action
onset and the outcome. The latencies indicated correspond to the median and confidence
interval of the position of the windows in which a change was detected.
Task event Activation
latency
Peak
latency
Reward modulation
latency
Delay modulation
latency
Cue onset 0
(-25; 75)
75
( 50; 175)
150
(125; 225)
350
(250; 450)
Bar release -250
(-325;-200)
-100
(-125;-75)
-512.5
(-612.5; -143.75)
-187.5
(-400; -150)
Outcome 0
(-25; 25)
125
(75; 150)
200
(162.5; 375)
100
(100; 100)
At the cue onset, the modulation of LC activity as a function of expected Reward (Figure
43A&C) and Delay (Figure 43B&C) levels appeared later than the average change in firing rate.
Indeed, the average latency of Reward and Delay effects were significantly longer than the
average peak latency (Wilcoxon signed rank test with continuity correction: V=276, p<0.001
and V=70.5, p<0.001; for Reward and Delay, respectively). In addition, the timing of the
Reward and Delay modulation differed in time, since the average latency of the Reward
modulation was significantly shorter than that of the Delay modulation (V=78.5; p<0.001). This
suggests that the source of LC inputs responsible from the global activation after cue onset
might differ, at least in time, from those responsible for Reward and Delay modulation.
Experimental results
161
Figure 43: Summary of the dynamics of the modulation of rate by task parameters (Reward
and Delay) around cue onset (A,B,C), action onset (D, E, F) and the outcome (G, H, I).
(A, B, D, E, G, H) Plot of the betas for Reward (A, D, G) and Delay (B, E, H) in the GLM for spike
count across time at cue onset (A, B), action onset (D, E) and the outcome (G, H), calculated
via a sliding window procedure (window size=200ms, step=25ms). One line corresponds to the
betas for one individual neuron at each time point (middle of the window of analysis). The color
scale is displayed below the graphs. The green dots represent the latency to the onset of the
effects.
(C, F, H) Plot of the mean rate and of the mean of the betas for Reward (green) and Delay (blue)
across neurons across time around cue onset (C), action onset (F), and the outcome (H).
At cue onset, Reward had a positive effect on rate (A,C) and Delay had a negative effect on
rate (A, C). (C) The latency to the peak in rate was shorter than the latency of the effect of
Reward which was shorter than that of the effects of Delay.
Before action onset, Reward had a small negative effect on rate (D,F) and Delay had a positive
effect on rate (E, F). (F) The effects of Reward and Delay appeared earlier in time than the peak
in rate.
At the outcome, Reward had a small negative effect on rate (G,I) and Delay had a strong
positive effect on rate (H, I). (I) The latency to the peak in rate and to the beginning of Delay
effects were not significantly different while the latency to Reward effects was greater than
the latency to the peak in rate.
At the bar release, the latencies of the Reward (Figure 43D&F) and Delay effects (Figure
43E&F) were indistinguishable (p>0.05). In addition, the effects of both Reward and Delay
appeared earlier than the peak latency of the firing rate modulation (Reward: V=16; p=0.0016;
Delay: V=23; p<0.001) but their latency was comparable to that of the onset of the rate effects
(p>0.05 for both Reward and Delay). In other words, the dynamics of the activation of LC
neurons when monkeys initiated the bar release was similar to the dynamics of the
modulation by Reward and Delay factors.
At the trial outcome, the Delay modulation (Figure 43H&I) appeared after the
beginning of the increase in rate (V=1178; p<0.001) and I could not find a difference in
latencies between the beginning of Delay effects and of the peak in rate (p>0.005), indicating
Experimental results
162
that these may be simultaneous. The effects of Reward (Figure 43G&I) began significantly later
that the peak in rate (V=188, p<0.001), but this effect was very small.
In summary, the complex dynamics of cue-evoked activity indicates that LC neurons
received an initial non-specific excitatory input, followed first by a positive modulation by
Reward information, and later by a negative modulation by Delay information. By contrast, at
the bar release and the trial outcome, the increase in firing rate occurred at the same time as
the positive modulation by Delay information and the negative modulation by Reward
information.
c) Intrinsic neuronal excitability shaped the activation
in response to task events and the encoding of task
parameters and behavioral variables
(1) Higher baseline was associated to higher peak of activity
Total and colleagues (2018) have reported a certain level of heterogeneity across LC
neurons regarding their intrinsic excitability in vivo, and such differences were found to relate
to different projection patterns, and therefore differential implications in cognitive processes
(Uematsu et al., 2017, 2015). In our dataset, I looked for heterogeneity in baseline rate across
neurons, and related differences in baseline rate to differences in activation strength around
task events of interest (cue, action, outcome), and to differences in task parameter encoding
strength.
Even if the response pattern of the population of LC neurons was relatively
homogenous (Figure 41), response magnitude to each of the task events varied across the 75
recorded neurons. I reasoned that if the variability in response magnitude across neurons was
simply related to their global excitability, the magnitude of evoked responses should be
predicted by the neuron baseline activity, across the population (Figure 44A,E,I).
In practice, for each neuron, I identified the moment at which rate was maximal around
the onset of the three task events, using a sliding window procedure or spike counting (see
methods of Bornert and Bouret, in prep). I computed the mean spike count of each neuron at
the peak of its firing rate (i.e. in the time window of the sliding window procedure with the
highest mean spike count), for all three task events, to evaluate the strength of the excitatory
responses. For each neuron, I evaluated baseline activity using spike counts before cue onset
Experimental results
163
([-300;0ms] epoch before cue onset). After computing the strength of the excitatory responses
for each neuron, I evaluated their correlation with baseline rate across the population of all
75 LC units. Note that similar results were obtained if the window used for evaluating baseline
rate was systematically taken before the activation in response to the event of interest, rather
than taking the same reference for all events in the inter-trial interval. For simplicity, results
reported below use a single mean baseline rate value per neuron, computed before cue onset.
Figure 44: Summary of the correlations between baseline rate and the strength of the response
to task events and of modulation of rate by task parameters around task events.
(A, E, I) Correlations between baseline rate (pre-cue rate) and post cue (A), pre-action (E) and
post outcome (I) peak rate. Baseline rate and peak rate are positively correlated.
Experimental results
164
(B, C, D, F, G, H, J, K) Correlations between baseline rate (pre-cue) and the maximum of the
absolute value of betas for Reward (B, F, J), Delay (C, G, K), willingness to work (D), and RT (H)
across time by neuron. For all regressors, the maximum of their effect size is positively
correlated to baseline rate.
(L) Correlation between maximum strength of effects of Reward and Delay on rate after cue
onset. The maximum strength of effect of Reward was positively correlated to the maximum
strength of Delay effects across neurons.
R: Reward; E: Delay; WTW: willingness to work
At cue onset, the rate at the peak was both positively correlated with the baseline rate
(R=0.826; t(73)=12.549; p<0.001). I also found positive correlations between pre-action peak
rate and baseline rate (R=0.740; t(73)=9.41; p<0.001) and between baseline rate and post
outcome peak rate (R=0.888; t(73)=16.514; p<0.001). These results imply that neurons with
higher baseline rates had higher firing rates during periods of event related activation.
(2) Higher baseline was associated to higher signal-to-noise
ratio
Classical theories of tonic versus phasic LC activity modes state that high tonic activity,
i.e. high LC baseline rate is generally associated to low signal to noise ratio in LC activity, i.e.
small differences between baseline rate and evoked responses, with the opposite pattern in
the phasic mode (Aston-Jones and Cohen, 2005). This theory focuses on within-neuron
changes in baseline firing. Here I question whether it can apply across neurons, i.e. whether
high-baseline neurons tend to have a smaller difference between baseline rate and rate at the
maximum of evoked activation. The opposite would be predicted from excitability accounts
of neuron-by-neuron excitability differences. If neurons that are more excitable have a higher
baseline rate, they should also respond more strongly to events. This analysis is
complementary to the one presented in the previous paragraph. Indeed, the direct correlation
between baseline rate and peak rate could have been driven by baseline differences, and
cannot exclude the possibility that beyond this baseline effect, neurons with a higher baseline
display a smaller increase in rate than those with a lower baseline. This is illustrated by Figure
45.
Experimental results
165
(A) (B)
Figure 45: Two alternatives for the relationship between baseline rate and signal-to noise ratio,
represented using schematic representations of the mean rate of virtual example units across
time around an event of interest.
(A) Excitability account: increased baseline leads to increased signal to noise ratio, i.e.
increased difference between peak rate and baseline. Peak rate (without subtracting baseline
rate) and baseline rate are positively correlated.
(B) Gain theory account: higher baseline corresponds to lower signal-to-noise ratio.
Nonetheless, peak rate and baseline rate are correlated.
I therefore examined whether there was a correlation between baseline rate and the
amplitude of the evoked change in firing from baseline. I measured the absolute value of the
difference between baseline rate and peak rate, and evaluated the strength of the correlation
between each of these values and baseline firing across neurons. The magnitude of excitatory
responses displayed a significant positive correlation with baseline rate across neurons
(respectively R=0.51; t(73)=5.1; p<0.001). This indicates that besides having generally higher
rates throughout the task, neurons with higher baseline rates responded to task events more
strongly, exhibiting stronger activations.
Rate Rate
High baseline
Middle
baseline
Low baseline
Time Time
Signal/noise Signal/noise
Experimental results
166
(3) Higher baseline was associated to stronger encoding of
task parameters and behavioral variables
Finally, I examined the relation between baseline rate and the encoding of task
parameters (Reward-Delay conditions, Figure 44B,C,F,G,J,K) as well as event-related
behavioral measures (willingness to work at the cue, Figure 44D, RT at the action, Figure 44H)
across this sample of LC neurons. In order to relate the effect sizes of Reward, Delay, RT and
willingness to work to baseline rate and to each other, I determined for each neuron the
window of the GLM for spike count sliding window analysis corresponding to the extremal
value (maximum of the absolute value) of the coefficients for each of these parameters. In
other words, I measured the relation between baseline rate and strength of encoding, by
taking the maximum of the absolute value of each regression coefficient. The results of this
analysis are reported in Table 2.
Table 2: Results for the Pearson’s correlations of baseline rate with effect sizes (maximum of
the absolute values of coefficients from the spike count GLM) for reward, delay, RT and
willingness to work, at cue onset, action onset and the outcome.
Correlation of
baseline rate
with …
… the maximum of
absolute beta
reward
… the maximum of
absolute beta delay
… the maximum of
absolute beta RT
… the maximum of
absolute beta WTW
Cue onset R = 0.75
t(72) = 9.71
p < 0.001
R = 0.79
t(72) = 10.86
p < 0.001
R = 0.27
t(67)=2.32
p=0.023
Action onset R = 0.56
t(72) = 5.84
p < 0.001
R = 0.64
t(72) = 11.65
p < 0.001
R = 0.69
t(72) = 4.50
p < 0.001
Outcome R = 0.54
t(72) = 5.40
p<0.001
R = 0.45
t(72) = 4.90
p < 0.001
Experimental results
167
Around all three studied task events, effect sizes i.e. encoding strengths were positively
correlated to baseline rate. This indicates that the variability in baseline rate across neurons
could account for some of the variability in sensitivity to task parameters and behavior.
Beyond confirming the previous result that neurons with higher baseline rates are more
sensitive to activating influences, this demonstrates that they are also providing a stronger
input about task events and behavioral responses.
(4) Encoding of Reward discounted value
Finally, I examined whether Reward discounted value, approximated by the difference
between Reward and Delay, was reliably encoded by the LC neurons at the time of the cue.
Specifically, I looked for a negative correlation between the neuron-by-neuron maximal
effects of Reward and of Delay (extrema of the regression coefficients across time around the
three task events, with its sign) (Figure 44I). I indeed found a significant negative correlation
(R=-0.75; t(73)=-9.56; p<0.001), yet this correlation could have been artificially created by the
positive correlations between regression coefficients and rate. Therefore, I then fit a GLM for
the neuron-by-neuron maximal effect of Delay, with the neuron-by-neuron maximal effect of
Reward and baseline rate as parameters. Over and above the effect of baseline rate (β=0.11;
p=0.012), the coefficient for the maximal effect of Reward on the maximal effect of Delay was
significantly negative (β=-0.55; p<0.001). Such results were not replicated at action onset and
the outcome (p>0.05 for both correlations).
3. Discussion of additional results and of study limitations
LC neurons displayed a transient response to the cues signaling the onset of a trial and
providing information about that trial (here, Reward and Delay). As in previous studies, the
strength of the response to cue onset was relatively modest, compared to the strong
activation at the bar release, but the magnitude of this response was strongly modulated by
task parameters (Bouret and Richmond, 2009; Varazzani et al., 2015). In line with previous
work, the intensity of the initial phase of the activation was equivalent across conditions, and
LC neurons only started to discriminate across reward conditions about 100ms after the onset
of the activation (Bouret and Richmond, 2015). Interestingly, the modulation by the
Experimental results
168
information about delay duration appeared even later (about 200ms after the beginning of
reward modulation). Although Varazzani and colleagues (2015) did not examine this directly
in their study on the effort/reward trade-off, the encoding of reward after cue onset did
emerge earlier than that of effort (Varazzani et al., 2015, fig. 5C).
The distinct timing of Reward- and Delay-related effects on activity suggests that the
corresponding modulation of LC activity at the cue could be triggered by separate inputs. But
since the dynamics with which such information is processed remains unclear, LC neurons may
still be contributing to the integration of these parameters, for instance assuming temporal
summation in target areas. This ability to integrate Reward and Delay inputs in a coherent
fashion may account for the relation between LC activity and willingness to work (also
described in Jahn et al., 2020). Indeed, since the effect of condition value, approximated using
engagement probabilities, on cue-evoked activity decreased when Reward and Delay were
added as covariates in the model, LC activity might very well play a role at the interface
between processing task-related information and modulating engagement.
At the time of the outcome of trials, after the delay, when the reward previously
announced by the cue was delivered, LC neurons also displayed a strong activation.
Noticeably, they strongly encoded the Delay category, positively, but not Reward size.
Conversely to what was observed at cue onset, this pattern excludes an encoding of the
discounted value of the Reward. Along the same lines, there is no reason to believe (but no
reason to exclude either) an interpretation in terms of prediction error. This increased LC
activation for longer delays could be related to the general level of salience of the outcome,
in relation to the level of surprise (Aston-Jones and Bloom, 1981; Bouret and Sara, 2004;
Vankov et al., 1995). Indeed, assuming that estimating the timing of the outcome becomes
more difficult as delay increases, longer delays could make reward delivery more surprising,
and therefore more salient. Again, further experiments would be necessary to clarify that
point.
In line with previous studies, the baseline firing rate of the recorded LC neurons
remained stable during the course of the experiments, as long as monkeys remained awake
(Bouret and Richmond, 2015, 2009; Bouret and Sara, 2004; Clayton et al., 2004; Kalwani et al.,
2014). Interestingly, the difference in baseline firing rate across LC neurons was strongly
Experimental results
169
predictive of the differences in response magnitude, both in terms of event-related changes
in firing rate (within trials) and in terms of parametric modulation by task parameters or
behavioral responses (across trials). Neurons displaying the highest firing rates also displayed
the strongest responses, suggesting that individual LC neurons could readily be defined in
terms of global excitability i.e. general susceptibility to excitatory inputs. These results are at
odds with the intuition that baseline activity is noise and has a negative influence on evoked
activity. In the light of the recent findings in rats by Chandler et al. (2014), who demonstrated
that LC neurons projecting to frontal cortices were more excitable than neurons projecting to
the motor cortex, it would be crucial to verify if this relationship is also valid in monkeys. If it
were the case, that would provide a way to identify LC neurons putatively projecting to the
frontal cortex based on their firing and assess other functional properties, such as the relation
with executive functions, and thereby further investigate the functional heterogeneity of the
LC.
More generally speaking, even if this study clearly lacks that level of anatomo-
functional characterization, these results clearly support the emerging idea that the LC is
functionally heterogeneous (e.g. Chandler et al., 2014; Schwarz and Luo, 2015). Moreover, LC
units seem to be grouped into distinct functional modules displaying context-dependent
activity (Likhtik and Johansen, 2019; Uematsu et al., 2017). In that frame, any
neurophysiological feature that could be used to characterize individual LC units in behaving
monkeys might be useful to assess the relation between groups of LC neurons (based on these
features) and behavioral/cognitive contexts and thereby provide a useful complement to
experiments conducted in rodents.
Despite such heterogeneity, LC neurons analyzed as a population seem to provide a
reliable encoding of both task parameters (Delay and Reward) and of behavioral parameters
(RT, willingness to work). For instance, around cue onset, the strength of the encoding of the
task parameters necessary for computing reward value (size and delay) were encoded in a
coherent fashion by LC neurons, ruling out the possibility that Delay and Reward information
were encoded by distinct neuronal populations, putatively sending axons to different brain
areas. Such a pattern is in contrast with what has been observed in the single-unit activity of
cortical areas believed to participate in value processing, such as the ventro-medial prefrontal
cortex (vmPFC, e.g. Hosokawa et al., 2013; Kennerley and Wallis, 2009b; San-Galli et al., 2018).
Experimental results
170
Despite providing valuable insight into the role of the LC in decision-making, and
especially control allocation into triggering costly actions, this study had several limitations.
First, decisions consisted in binary choices between engaging in trials or refusing them. Such
decisions could have been automatic instead of goal-directed. Moreover, placing cue
presentation during a period of waiting did not allow measuring the duration of decisions, at
least when the monkey decided to engage, as Jahn and colleagues (2020) have shown that
decisions to forego a trial can be timed using eye fixation breaks. These limitations are of
particular interest as LC neurons seemed to be activated by decisions to abandon trials (eye
fixation breaks, Jahn et al., 2020). Responses to cue onset could in fact be aligned onto
decisions whether to engage, and this could account for the relatively more variable latencies
of post-cues responses compared to per-action responses (see Fig. 3 in Bornert and Bouret, in
prep, and Figure 41A&C). The second study was designed in order to fill those gaps and
therefore complement previous results.
Experimental results
171
B. Study 2: LC neurons facilitate delayed effort
mobilization and monitoring in a two-alternative choice
task
1. Introduction
In the first study, we reported that LC neurons encoded the subjective difficulty of
triggering actions at the time of triggering them, and the difficulty of executing them while
executing them. We interpreted these modulations as a domain-general encoding of effort
production. Pharmacological studies pointed towards a beneficial effect of LC activity on effort
production, at least in the context of physical effort (Borderies et al., 2020; Jahn et al., 2018).
The dynamics of LC activity relative to effort production observed in our first study, combined
with the slow conduction velocity of LC axons, was incompatible with immediate
consequences on action-triggering and action execution. We suggested that LC activity around
effortful processes may have short-term (sub-second) but not immediate facilitative effects
on effort production, regardless of the nature of the challenge faced (physical or mental).
Moreover, based on studies showing a potential relation between LC activity and adjustments
of effort allocation (at least following errors, Murphy et al., 2016; Sellaro et al., 2015), we
proposed that LC activity may serve as a feedback signal regarding effort production costs and
consequences, causing adaptations of effort-allocating decisions. In order to test our
hypotheses, we recorded the activity of LC neurons in two-alternative effort-reward choice
tasks.
Besides confirming the findings regarding the domain-general encoding of effort by LC
neurons, our second study first aimed at testing the relation between LC activity around effort
production and later effort production. When presented with two alternatives differing in
terms of required effort and associated rewards, subjects choose between offers by making a
tradeoff between effort and reward. Consequently, if effort production was facilitated, the
effect of effort on binary choices should decrease, and vice versa, as was observed in Borderies
et al. (2020). The tasks that we designed required physical effort production (differential force
exertion on a grip). Physical effort facilitation should lead to an increased probability to choose
high-physical effort offers. Additionally, these paradigms allowed combining the assessment
of effort sensitivity of choices to an evaluation of the force exerted during the execution of
Experimental results
172
the chosen action. Within force conditions, increased exerted force should indicate facilitated
effort production, especially if it was combined to decreased effort sensitivity of choices.
Cognitive control exertion optimizes the decision process (Botvinick and Cohen, 2014;
Cohen et al., 1990; Missier et al., 2012). In that sense, in tasks offering overt decisions between
several alternatives such as our two-alternative choice tasks, if LC activity facilitated effort
production in general and therefore cognitive control exertion, we would expect that
following high LC activity, choices would be more accurate. Several studies have
complementarily demonstrated that cognitive control is crucial for decisions that go against a
bias (Epstein, 1994; Shenhav et al., 2013; Shiffrin and Schneider, 1977; Sloman, 1996).
Monkeys have a natural tendency towards avoiding offers requiring high effort production, as
shown by trial omissions in one-option tasks and two-alternative choice tasks (respectively:
Jahn et al., 2020; Varazzani et al., 2015; and Borderies et al., 2020), and binary choices in two-
alternative choice tasks (Borderies et al., 2020). In that sense, facilitated cognitive control
exertion could alternatively cause decreases in the sensitivity of choices to effort, i.e. increase
the probability to choose high-effort options.
Second, this study aimed at exploring the role of LC activity, especially its modulations
by effort production, as a feedback for the effort allocating system. We theorized that LC
activity around the time of effort production in one trial would cause adjustments of decisions
to allocate effort in the next trial. Practically, in our two-alternative choice tasks, we expected
action-related LC activity to predict adjustments in effort-based binary choices in the next trial.
We expected that if LC neurons activated more around the action of trial N, signaling high
effort production, the next choice would have increased effort sensitivity (decreased
probability to choose the high-effort offer).
In a pilot study, we recorded LC neurons in a two-alternative choice task quite similar
to the one used in Borderies et al. (2020). It consisted in presenting two offers, i.e. cues
indicating two force-reward combinations, during a waiting period, and then asking the
monkeys to execute the chosen action to obtain the chosen reward. One monkey was trained
on that task, and two dozen LC neurons were recorded, providing some preliminary
indications of the pattern of activity of LC units in such tasks.
Experimental results
173
Our initial paradigm, however, required optimization for several reasons. First, the
monkey had to materialize his choices by pressing a grip located on the side of presentation
of the chosen cue at the required force. Consequently, in each trial, the monkey could either
choose to keep his hands on the same grip as the one used in the previous trial or choose to
switch. He developed a strong bias towards keeping his hands on the same grip, thus
frequently ignoring cue information to make his choice.
As in the first study (Bornert and Bouret, in prep), we planned on using RT as a measure
of the subjective difficulty of triggering actions. Since our task involved a choice between
alternatives, RTs should also vary with choice difficulty and probably with the cognitive control
exerted to make the choice (Hunt et al., 2012; Krajbich et al., 2010; Krajbich and Rangel, 2011;
Milosavljevic et al., 2010; Morris et al., 2019; Philiastides and Ratcliff, 2013). These effects
should be additive (Dmochowski and Norcia, 2015). However, in the pilot task, RTs were hardly
interpretable: in addition to effects of side of press (postural effects), RTs varied depending
on the previous hand position (previous grip pressed) relative to the current grip chosen. For
instance, if the monkey chose the left side option at trial N, he was much faster if he had
already used that grip at trial N-1, since he did not need to change hand position, compared
to trials in which he had used the right side grip at trial N-1. Additionally, the cues were
presented during a waiting period, which did not allow decision duration to be measured.
Finally, regarding willingness to work, since the task did not impose eye fixation (the
monkey had to examine both options, so gaze fixation was not an option), it was impossible
to assess whether the monkey looked at the cues and decided to refuse to engage, or did not
even look at the cues.
Consequently, we conceived a new paradigm, fixing the imperfections of the previous
one (see materials and methods, part II.B.2). Practically, monkeys were first asked to press on
a middle grip before being presented with a pair of offers, for which they had to press one of
the two side grips. This middle press had several functions: first, it removed the bias of the
monkey to press the same grip at trial N as he had at trial N-1. Second, in most trials, the
monkey would press the middle grip without knowing what the subsequent offers would be,
allowing us to know when he was willing to look at the cues, and made an actual decision
whether to engage in the rest of the trial. Third, since the LC is believed to participate in effort
production for action triggering, execution and action consequences monitoring, designing a
Experimental results
174
task with several steps allowed testing whether LC activity at the first step (middle press) could
influence decisions to engage and binary choices and action performance at the second step
(choice between offers and associated press). Additionally, in this new task, the cue was no
longer presented during a waiting period, i.e. the cue with the two offers was also the go-
signal and the monkey could trigger the chosen option as soon as he had reached a choice.
This would allow timing the duration of the RT consisting of the sum of the durations of the
action selection and the action triggering periods. Gaze position was recorded throughout the
trials and I designed an eye-tracking procedure allowing identification of the timing of the end
of the decision process (see materials and methods, part II.B.2).
2. Materials and methods
a) Animals
One rhesus macaque (Macaca mulatta), monkey J (male, 12kg), was included in this
study. During testing days, the monkey received water as reward, and on non-testing days, he
received amounts of water matching his physiological needs. The experimental procedures
were designed in association with the veterinarians of the ICM (Brain and Spine institute),
approved by the Regional Ethical Committee for Animal Experiment (CREEA IDF n°3) and
performed in compliance with the European Community Council Directives (86/609/EEC).
b) Behavioral paradigm
The monkey was extensively trained on a two-alternative forced choice task (Figure
46A). During sessions, he squatted in a primate chair, in front of a computer screen. The
monkey was head restrained, to allow eye-tracking and electrophysiological recordings. Three
force-sensitive grips (middle, left and right grips, M2E, France) were installed on the anterior
part of the chair, within reach from the hands of the monkey. A metal straw delivering water
rewards was placed between his lips.
Trials begun with the presentation of a central red dot (wait-signal) of variable duration
(100±50ms), after which it turned green and remained green for 3000ms maximum. During
this period, the monkey could press on the middle grip, and any small force exertion was
sufficient. If the monkey pressed the grip, the dot turned blue, and remained blue for a
variable duration (1000±100ms). This duration was imposed to be able differentiate middle
Experimental results
175
action-related LC activity from cues onset related activity. After that delay, the central blue
dot disappeared and two cues appeared on the sides of the screen, one on the right side and
one of the left side, as well as one green dot per side. No waiting period was imposed following
cues presentation, and the monkey could virtually immediately trigger the chosen action for
the whole duration of cues presentation (2250±250ms). The action consisted in pressing the
grip located on the side corresponding to the side of presentation of the chosen cue at the
level of force instructed by the cue. Once a press had been detected on one grip, the monkey
had 1000ms to reach the force threshold imposed by the condition. Levels of force had been
calibrated through training to identify combinations that did affect choices but did not
demotivate the monkey. If the force produced reached the correct threshold of force, the
green dot on the chosen side turned blue, and the force had to be maintained for an additional
80ms. Inter-trial intervals were 475±25ms long.
Each cue consisted of a reward cue (number of blue drops) and a force cue (rectangle
of variable size, Figure 46B). Reward and Force were orthogonalized and left and right offers
were picked from a pseudorandom distribution.
(A) (B)
Figure 46: Two-alternative choice task.
(A) The task consisted in two steps: pressing a middle grip, after what two offers were
presented, and choosing between these options. Choice was materialized by pressing on the
grip located on the same side as the chosen option.
(B) Offers consisted in a combination of Reward (1,2 or 4 drops of water) and requested Force
(1,2 or 3). Each combination was associated to a specific cue, which the monkey had learned
to use.
Erroneous press
Omission
Experimental results
176
Failure to produce the requested force or to sufficiently maintain it (erroneous press),
failure to produce a press on the middle grip at the first step, or on any side grip at the second
step (omission), and presses on any grip during the red dot presentation led to the trial being
restarted (from the middle press step). Consequently, the same offers were presented at the
second step. If the monkey had reached the cue presentation in the failed trial, he already
knew what offers he would get, and the cues therefore provided no new information
(repeated trial). If he had made no error at the previous stage or made an error that did not
allow him to see the cues (omission of middle grip press), the cues carried novel information
(non-repeated trial).
c) MRI, surgery and recordings
A 3T MRI of the monkey’s brain was performed under general anesthesia to determine
the position of the LC, based on previously described anatomical criteria (see part I.B.1.a) for
review and Bornert and Bouret, in prep, materials and methods for extensive description). A
few days after the MRI, a headpost and two recording chambers, each located on top of the
theoretical position of the LC in each hemisphere, were placed on the skull of the monkey
during a sterile surgical procedure under general anesthesia. In this study, only the right side
chamber was used, the second had been placed in prevision of bilateral DREADDs injections.
The right side chamber had a 14° angle to the vertical plan. After a few weeks of recuperation,
a sterile craniotomy was performed in the right side chamber, preserving the dura, to allow
electrode descent in that chamber.
Electrophysiological methods were the same as described in Bornert and Bouret, in
prep (materials and methods) and in several other papers (Bouret et al., 2012; Bouret and
Richmond, 2015, 2009; Jahn et al., 2020; Varazzani et al., 2015). In brief, before each recording
session, I cleaned the recording chamber, and placed a grid with holes 1mm apart in the
chamber. I introduced a metal guide tube (CRIST, 49mm long) in the grid hole that was
identified as positioned on top of the body of the LC in the right hemisphere (Figure 47A). In
that tube, I inserted a tungsten micro-electrode (UEWLEHSM3PNM, FHC, Bowdoin, ME). The
electrode was then lowered using an oil hydraulic Microdrive (Narishige).
Experimental results
177
(A)
(B)
Figure 47: Electrophysiological recordings: location in the brain and pharmacological
validation.
(A) MRI of the brain of Monkey J (T2 contrast, visualized in Brainsight software). The grey
cylinders on top of the skull are the recording chambers. In the right side chamber (right side),
a grid had been placed (grey stripes in chamber). The trajectory used for the recordings is
shown in green. The LC was approximately located 63mm below the top of the grid.
(B) Discharge of example LC unit after clonidine injection (20μg/kg, IM, green arrow). After 5
minutes (time in x-axis), rate (y-axis) started decreasing. The monkey was drowsy (red arrow)
after about 15 minutes, and LC neuron rate dropped after 20 minutes. During the period of
Rat
e (H
z)
3
6
Time (s)
Injection
Drowsy
Opened eyes
* *
* *
* *
Experimental results
178
drowsiness, LC activity increased just before increases in arousal (eyes opening, purple stars).
After about an hour, LC activity increased again and the monkey woke up shortly after.
During the descent, the pattern of discharge and of response to events and stimuli
(sounds, touch, movement) of encountered neurons were used in combination of the MRI
scan to determine the position of the tip of the electrode in the brain. LC neurons were
identified using extensively described electrophysiological criteria (low rate of spontaneous
activity (<4Hz), broad waveforms (>0.6ms until the first peak), and characteristic activation-
pause response to brief auditory or tactile stimuli (e.g. calling monkey’s name, knocking on
the door, or clapping hands). I confirmed this identification by performing clonidine injections
(15-20μg/kg, IM, SIGMA). Under clonidine, the monkey became drowsy. Drowsiness preceded
a decrease the activity of concomitantly recorded LC neurons (Figure 47B). Once LC activity
had become low, after about 20 minutes, increases in activity preceded periods of decreased
drowsiness (eye opening, small movements). The activity of neurons increased before the
monkey woke up.
d) Data analysis
All analyses were performed using the Matlab software. For all analyses, offer
parameters (Rewards, Forces) and behavioral parameters (engagement rate, response time,
exerted force) were z-scored before fitting models.
The procedure for RT determination for middle presses and side presses was identical
to the procedure described in the first study (Bornert and Bouret, in prep), presented above.
The time courses of the force exertion on each of the three grips with the RTs determined by
this method are shown in Figure 48.
The procedure for rate assessment across time (sliding window procedure) was the
same as the one used for the first study (Bornert and Bouret, in prep).
(1) Eye-tracking and determination of action-selection and
action-triggering time
In this task, gaze position (position of the right eye) was recorded in 32/34 sessions i.e.
for 50/54 recorded neurons (ISCAN system). I used the signal-detection task of Study 1, in
which the monkey had to detect that small dots switched from red to green and trigger a
Experimental results
179
press, to calibrate the relationship between the recorded signal and actual gaze position.
Practically, considering the small size of the dots and the unpredictable timing of the color
switch, I made the postulate that the monkey was necessarily looking at the dot when he
triggered his response if the response was correctly timed. Using trigonometry, I then inferred
the relation between pupil center vertical and horizontal eccentricity recorded by the eye-
tracking system and vertical and horizontal angles from screen center. Consequently, I could
determine within which intervals the recorded horizontal and vertical eccentricity of pupil
center should be to ensure that the monkey was looking at the left or right side offer. Note
that identical results were obtained by simply determining the position of the middle of the
screen (via calibration or by taking the mean of the horizontal eye position signal throughout
the whole session) and making the approximation that the monkey was looking at the right
side offer when he looked to the right of the screen center, and vice versa.
Figure 48: RT determination and decomposition: forces exerted in the two-alternative choice
task and eye-tracking procedure.
(D) Reaction Time (Go-signal to press)
Gaze alternates between options: action-selection time
Gaze locked on chosen option: action-triggering time
Action triggered
Right side press
G
O
+
C
U
E
S
Experimental results
180
(A-C) Time course of force exertion (baseline subtracted) after middle press onset (pink vertical
line) or side press onset (orange vertical line), on (A) the middle grip, (B) the left side grip and
(C) the right side grip. All sessions were pooled. The thickness of the lines of the plot
corresponds to one SEM above and below the mean exerted force. For the two side grips (B,C),
trials are separated according to the Force condition (force requirement).
(D) Schematic representation of the procedure based on eye-tracking data used to decompose
RT into action-selection and action-triggering time. After the Go-signal (green dot, here
represented as a green rectangle), during RT (red arrow), the monkey (brown ellipse, eyes as
black circles) usually looked alternatively at the two offers presented on screen (grey
rectangles with cues) for a certain amount of time (dashed orange line, gaze direction
extracted from right eye position represented by blue lines). Then, he locked his gaze onto the
option that he would choose (execute) until actually triggering the action (dotted orange line).
We refer to the time between cues onset and gaze locking onto the cue of the chosen offer as
“action-selection time” and to the time between gaze locking onto the cue of the chosen option
as “action-triggering time”.
To decompose RT, after determining it from the force exertion profiles, I first identified
periods of times during which the monkey looked at either cue during RTs. I made the
postulate that the monkey was making his choice (action-selection) for as long as he
alternately gazed at both offers. Then, when his gaze stopped alternating and locked onto one
offer, the monkey would be triggering the action he chose to execute (Figure 48D).
Consequently, I identified the timing of the last saccade onto the chosen option. The time
from cues onset to this saccade corresponded to action-selection time, while the time from
this saccade to action-triggering (detection of the beginning of the press) corresponded to
action-triggering time. One limitation of this technique comes from outlier trials, in which the
monkey either did not look at both cues (kept his gaze onto one cue until he responded), or
did not look at the chosen option when he triggered its execution. These trials were excluded
from analyses requiring the use of action-selection and action-triggering times.
(2) Value inference from choices
I used the choices between offers made by the monkey throughout sessions to
determine the mean subjective value of each offer i.e. combination of Reward and Force for
Experimental results
181
each session. Session-by-session, I fit a logistic regression for choices (0 if the monkey chose
the left side offer, 1 if he chose the right side offer) with right side Forces and Rewards minus
the left side Forces and Rewards as parameters. Indeed, considering the polarization of the
logistic regression, weights for left side offers had to have a change of sign:
Choice ~ logit ( c + βRr . RR + βFr . FR - βRl . RL - βFl . FL )
With c a constant term, RR the Reward of the right side offer, FR the Force of the right side
offer, RL the Reward of the left side offer, FL the Force of the left side offer, and β the associated
coefficients in the logistic regression for choices.
All the trials in which the monkey made a side press were included in the regressions.
For each session, the mean weight of the Forces and of the Rewards on subjective values was
thus determined (β in the previous formula). For each trial, I then inferred the value of the
offers from these weights:
Voffer LEFT = βRl . Roffer + βFl . Foffer
Voffer RIGHT= βRr . Roffer + βFr . Foffer
With Voffer LEFT and Voffer RIGHT the subjective value for the considered session of one specific
offer presented on the left or right side, respectively, as a function of its Reward (Roffer) and
Force (Foffer) characteristics and of the coefficients for Reward and Force on the corresponding
side (left side, respectively βRl and βFl, right side, respectively βRr and βFr).
3. Results
a) Behavior
One monkey was trained on a two-alternative choice task consisting of two steps (see
part II.B.2.b) and Figure 46 for details). The first was pressing on a middle grip, after what two
offers, a combination of Reward size and requested Force, were presented. The second step
was making a choice between the two offers and executing the chosen action, by pressing the
grip placed on the side corresponding to the side of presentation of the chosen offer.
Erroneous trials (pressed at the wrong time, incorrect pressing force, or omission) were
repeated. This means that, if the trial was aborted (omission) or failed (erroneous press) after
cue presentation, the cues had already been seen by the monkey, and until the trial was
Experimental results
182
completed, subsequent cue presentations carried no novel information. Behavior was
analyzed for the 34 recording sessions.
Throughout the analyses of the behavior, I looked for indications that the monkey
invested effort in the task, to decide to engage in actions, choose between offers, trigger
actions, and execute them, and examined the relationship between the different effortful
processes. Regarding cognitive control exertion to engage in actions and action-triggering, I
expected that the time taken to trigger the actions would scale negatively with the tendency
to disengage from the task and would increase throughout sessions. This would indicate that
this time could serve as a measure of the subjective difficulty of triggering actions, and possibly
of the cognitive control exerted.
Regarding binary choices between effort-reward combinations, I expected either one
of two possibilities. The time taken to make choices has been shown to reflect the cognitive
control invested in the decision (Hunt et al., 2012; Krajbich et al., 2010; Krajbich and Rangel,
2011; Milosavljevic et al., 2010; Morris et al., 2019; Philiastides and Ratcliff, 2013). In choice
tasks, behavior usually follows a speed-accuracy tradeoff: over and above choice difficulty, if
subjects take longer to decide, their decisions are usually more accurate (Chittka et al., 2009,
2003; Förster et al., 2003; Fudenberg et al., 2018). I looked for signatures of a speed-accuracy
trade-off in our task, i.e. for evidence that longer action-selection time was associated to more
accurate choices, via increased cognitive control exertion. Alternatively, cognitive control
exertion for the binary choice process may have reduced the strength of decision biases
(Epstein, 1994; Shenhav et al., 2013; Shiffrin and Schneider, 1977; Sloman, 1996). In that
frame, I would have expected that increased time spent on action-selection would lead to an
increased probability to choose high-effort offers, i.e. a decreased bias towards avoiding
them.
Regarding action execution, we defined effort as the process that mediated between
difficulty and performance. In other words, increased effort increases performance. Our task
inposed differential force execution on a grip to obtain a reward. Within force categories,
increasing force exertion increases the probability of success. In that sense, exerting effort to
execute actions should increase force exertion especially in high-force conditions. We
questioned the relation between physical effort production for action execution and cognitive
control exertion for engaging in the actions and triggering them. I expected that if engaging
in the actions and triggering them was subjectively more difficult, executing the action would
Experimental results
183
also be more difficult, and that less physical effort would be produced, resulting in decreased
force exertion and performance.
(1) Engagement in costly actions
(a) RT to the middle press represented the cost of
triggering the middle action and starting a trial
First, I analyzed behavior associated to triggering the middle press: engagement
(decisions whether to press the grip) and RT (time from go-signal to press onset). First, I aimed
at verifying the validity of RT as a proxy for the subjective difficulty of triggering the press on
the middle grip to start a trial. The distribution of middle press RTs, across all sessions, is
shown in Figure 49A. If middle press RT scaled with the difficulty to engage, it should have
increased throughout the session, as the monkey got tired, and in periods of time when the
monkey was less willing to engage.
To confirm that the cost of engaging in the middle press increased throughout the
session, I first verified that engagement in the middle press decreased throughout the session.
I pooled all the trials that were not repeated (in which the monkey did not see the cue before,
to avoid any potential value-based effects) of all the sessions and fit a logistic regression for
engagement with only trial number as parameter. The effect of trial number was significantly
negative (β=-0.15; p<10-144, Figure 49B), indicating that triggering a new trial was probably
more costly as the session went by. Consequently, if RT increased with increased cost to
trigger the middle action, I expected RT to increase throughout the session and to increase if
the current engagement rate was lower.
I then pooled all the trials in which the monkey pressed the middle grip during the
display of the go-signal, that were non-repeated trials. First, I fit a GLM for middle press RT
with only trial number as parameter. The effect of trial number was significantly positive
(β=0.11; p<10-22, Figure 49C). Second, I split sessions in halves (beginning and end), and
computed the engagement rate in the middle press in non-repeated trials for each half. Each
trial was therefore associated to the probability of engagement in its half of the session. I then
pooled all the non-repeated trials in which the monkey pressed the middle grip from all the
sessions, and fit a GLM for middle press RT with only block-wise engagement rate as
Experimental results
184
parameter. There was a negative effect of block-wise engagement rate on middle press RT
(β=-0.028; p=0.014, Figure 49D). These results provide two indications that RT to the middle
press could readily be used as a measure of the cost of triggering the middle action and
starting a trial in this task.
Figure 49: Behavior associated to the middle press was compatible with a positive relationship
between RT and the cost of triggering the middle press.
(A) Density function of middle press RT (bandwidth = 10). Most RTs are below 1 second.
(B) Logistic regression for engagement (whether the monkey pressed on the middle grip) with
trial number as parameter. Trial number had a negative effect on engagement, indicating that
the cost of engaging increased throughout the session.
Experimental results
185
(C,D) GLM for middle press RT with (C) trial number as parameter or (D) block-wise
engagement rate. (C) Trial number had a positive effect on RT, while (D) current engagement
rate had a negative effect, indicating that higher costs of triggering the action led to increased
RT.
*: p<0.05, ***: p<0.001, ctt: constant.
Overall, our results indicate that middle press RT scales with the subjective difficulty to
trigger the middle press, and therefore initiate new trials. Given that I focused on trials for
which the monkey did not know about the cues, this effect cannot be accounted for by effects
of the values of the offers from the choices.
(b) Decisions to engage based on cue information
Second, I focused on the second step requiring engagement: the two-alternative
choice. This novel paradigm allowed clearly identifying trials in which the monkey intended to
look at the cues (offers), and pressed the middle grip. After cues were presented, the monkey
could decide to make a choice (pick any option by pressing on any side grip) or to abort the
trial (no press, omission). I define choice engagement as the decisions whether to choose any
option offered (versus do nothing).
(i) Choice engagement increased with
increased sum of Rewards of options
I first studied the influence of offer parameters (offered Rewards and Forces) on choice
engagement in trials, trial-by-trial. In that aim, I compared the fit of different logistic
regressions for choice engagement (0, no press on side, 1, press on any side). Obviously, only
trials in which the monkey had made the middle press and therefore visualized the cues were
considered. Non-repeated trials from all the sessions in which the monkey pressed on the
middle grip and in which cues appeared were pooled. I expected Forces to have a negative
effect on choice engagement and Rewards to have a positive effect on choice engagement,
i.e. for the monkey to engage more in choices with high-value options. I compared three
models: a model with Rewards (left and right separately) and Forces (left and right separately)
as parameters, a model with Force minus Reward (left and right, separately), and a model with
the sum of Forces (left + right) and the sum of Rewards (left + right). I then compared their
Experimental results
186
Bayesian Information Criteria (BIC). The model including the sum of Forces and sum of
Rewards had the lowest BIC (BIC=1562.3), compared to the model including all the parameters
separately (BIC=1576.3) and to the model including the difference between Forces and
between Rewards on both sides (BIC=1574.7). The sum of the Rewards had a significant
positive effect on choice engagement on a trial-by-trial basis (β=0.71; p<10-13), and the sum of
Forces did not have any effect on choice engagement (β=-0.089; p=0.25, Figure 50A).
Figure 50: Decisions to engage in the choice in the two-alternative choice task.
(A) Coefficients in the logistic regression for choice engagement. Besides a strong bias towards
engaging, materialized by the significant constant (ctt) term, choice engagement was affected
positively by the sum of Rewards offered.
(B) Coefficients in the logistic regression for choice engagement, with RT of the middle press
added. Middle press RT had a significant negative effect on choice engagement.
ctt: constant term ; F: Force; R: Reward; *** : p<0.001; n.s.: non-significant, error bars
represent SEM.
Considering that these results suggest that the absolute weight of the sum of Rewards
on choice engagement was stronger than that of the sum of Forces, I then performed the same
logistic regression on trial-by-trial choice engagement, but session-by-session. Two sessions
had to be removed for that analysis due to a low proportion of trials in which the monkey did
not engage in choices. I compared the magnitude (absolute value of coefficients) for the sum
of Forces and the sum of Rewards, across sessions, using a paired t-test. This analysis
demonstrated that the magnitude of the effects of the sum of Rewards was larger than that
***
Experimental results
187
of the effects of the sum of Forces (t(31)=-3.68; p=0.001). In other words, the monkey placed
more weight on the total benefits (Rewards) offered in the trial than in the total costs (Forces)
when deciding to engage in choices.
Overall, these results indicate that the monkey engaged in trials more often when the
overall value, and especially the overall Rewards, of the offers were high, suggesting that
decreased overall value of offers increased the difficulty of engaging in the choice.
(i) Action-triggering time scaled with
difficulty to engage in the choice
Second, I examined the modulations of the time that the monkey took to trigger the
chosen action, and its relationship with the difficulty to engage in the choice. In this paradigm,
I recorded gaze position throughout trials. Using a calibration procedure, after cues onset, I
were able to identify periods of times during which the monkey looked at the right or left side
cue (see methods in part II.B.2.d)(1)). I define RT as the time elapsed between cues onset
(associated to the go-signal) and action onset (determined by finding the moment when the
force signal slope started changing, see procedure described in Bornert and Bouret, in prep,
materials and methods). Based on the literature on subcomponents of RT (Dmochowski and
Norcia, 2015), I made the hypothesis that RT was decomposable into two subparts: an action-
selection period and an action-triggering period, including the motor preparation (hand
positioning). I used eye-tracking to decompose RT into those two parts. After a period during
which gaze position alternated between the two offers (action-selection time), in most trials,
the monkey locked his gaze onto the chosen option. I refer to the period between gaze locking
and the beginning of the press on the chosen grip as the action-preparation period. The
duration of this period was highly correlated, negatively, to the duration of the action-
selection period, over and above effects of offer parameters and side of choice. For trials in
which the monkey never stopped looking at either option, or was not looking at the chosen
option while executing it, RT could not be decomposed into subcomponents. Additionally,
eye-tracking data was only available for 32/34 sessions. For the rest of this manuscript, every
analysis using action-triggering time and/or action-selection time therefore implicitly uses
Experimental results
188
only sessions and trials in which these two times could be determined. The distribution of RTs
and of their two subcomponents is shown in Figure 51A.
Figure 51: Action-triggering time increased with difficulty to engage in the choice
(A) Density plots of RTs (full line), action-selection times (dashed line) and action-triggering
time (dotted line).
(B) Coefficients in the GLM for action-triggering times. The chosen Reward had a negative
effect, while the chosen Force and the unchosen Force and Reward had positive effects. There
was a negative effect of action-selection time, meaning that the longer the monkey took to
decide on a choice, the faster he would be to trigger the chosen action. Additionally, Side had
Experimental results
189
a positive effect, indicating a longer duration of action-triggering time when the right side was
chosen.
(C) Coefficients in the GLM for action-triggering times. Besides the effects of offer parameters,
side and action-selection time, the choice engagement rate in the current trial block had a
negative effect on action-triggering time, indicating that increased difficulty to engage in the
choices led to increased time to trigger the chosen action.
ctt: constant term ; F: Force; R: Reward; *** : p<0.001; ** : p<0.01; * : <0.05 ; n.s.: non-
significant, error bars represent standard error of the mean (SEM).
I expected action-triggering time to increase with the difficulty of triggering the chosen
action. In order to examine that, I first fit a GLM for action-triggering time with task
parameters (chosen and unchosen Forces and Rewards), Side of press, and action-selection
period. All the trials in which the monkey made a choice, had looked at both cues, looked at
the chosen option at the time of the action, and in which the cue was unknown until
presentation (non-repeated trials) were included. The effect of the duration of the action-
selection period on the action-triggering period duration was significantly negative (β=-0.77;
p<10-100, Figure 51B), indicating that the monkey was faster to trigger the chosen action when
he had taken longer to choose between offers. Apart from that, the chosen Reward had a
quite strong negative effect (β=-0.23; p<10-119), and the chosen Force had a small positive
effect (β=0.028; p=0.0018). The unchosen Force and the unchosen Reward had a positive
effects (Force: β=0.030; p=0.0017, Reward: β=0.051; p<10-7), as well as Side (longer duration
for right side choices, β=0.089; p<10-6).
In order to confirm a link between action-triggering time and difficulty to engage in the
choice, I tested for a negative relationship between current engagement rate in choices and
action-triggering time. In an identical manner as what was done for middle press RT and
middle press engagement rate, sessions were cut in halves and choice engagement rate in
each half was determined. Each trial was therefore associated to a certain engagement rate
in its half of the session). Trials from all the sessions (same selection as previous GLM for
action-triggering time) were pooled. I performed the same GLM for action-triggering time as
before (with offer parameters, action-selection time and Side of press) but added block-wise
choice engagement rate. The effect of current choice engagement rate was significantly
Experimental results
190
negative (β=-0.044; p<10-6, Figure 51C), over and above the effects of offer parameters and of
side10. This indicates that the difficulty to engage in the choice scaled with action-triggering
time negatively, i.e. the monkey triggered the chosen action faster when deciding to engage
in the choice was less difficult.
(ii) Difficulty to engage in the choices scaled
with difficulty to engage in the middle press
Next, I tested whether the cost of triggering the action initiating the trial (middle press)
was related to the cost of engaging in the choice, over and above effects of task parameters.
Practically, I first examined whether current middle press engagement rate (within half of
session, see above for details) was correlated to current choice engagement rate. I pooled all
the non-repeated trials in which cues appeared and fit a logistic regression for choice
engagement with the sum of offered Forces, the sum of offered Rewards, and current middle
press engagement rate. Besides identical effects of the sum of Forces and of Rewards as
observed before (respectively β=-0.087; p=0.26; β=0.71; p<10-13), there was a positive effect
of current middle press engagement rate on choice engagement (β=0.16; p=0.043, Figure
52A). This provides evidence for the positive relationship between the difficulty to engage in
the middle press and in the choice.
Additionally, I expected middle press RT to have a positive effect on choice
engagement over and above effects of task parameters. I fit a logistic regression for choice
engagement (1 if the monkey made a choice i.e. pressed on one side grip, 0 otherwise) with
the sum of offered Forces, the sum of offered Rewards, and middle press RT as parameters.
Besides identical effects of the sum of Forces and of Rewards (respectively β=-0.087; p=0.27;
β=0.72; p<10-13), there was a negative effect of middle press RT (β=-0.25; p<10-4, Figure 52B)
on choice engagement. In other words, if the monkey perceived engaging in the middle press
as more difficult, i.e. was less willing to engage in it and spent more time triggering it, he also
had a lower probability to engage in the choice.
10 Effects of offer parameters, side and action-selection time remained unchanged: Chosen Reward: β=-0.23; p<10-120; Chosen Force: β=0.031; p=0.00062; Unchosen Reward: β=0.032; p=0.00099; Unchosen Force: β=0.049; p<10-7; Side: β=0.106; p<10-8; Action-selection time: β=-0.78; p<10-150.
Experimental results
191
Figure 52: The subjective difficulty to engage in the choice scaled with the subjective difficulty
to engage in the middle press.
(A,B) Logistic regressions for choice engagement. Over and above effects of the sum of Forces
and of Reward offered, (A) the current engagement rate in the middle press (engagement rate
in the current half of the session) had a positive influence on choice engagement, and (B)
middle press RT had a negative effect. This indicates that if the monkey was less willing or had
taken more time to trigger the middle press, he would be less willing to engage in the choice.
(C,D) GLMs for action-triggering time. Over and above effects of the sum of Forces and of
Rewards offered, (C) the current engagement rate in the middle press had a negative influence
on action-triggering time, and (B) middle press RT had a positive effect. This indicates that if
Experimental results
192
the monkey was less willing or had taken more time to trigger the middle press, he would take
more time to trigger the chosen action.
ctt: constant term ; F: Force; R: Reward; *** : p<0.001; * : <0.05 ; n.s.: non-significant, error
bars represent the SEM.
Next, I tested whether this relationship also applied to action-triggering time, i.e.
whether the monkey spent more time triggering the chosen action when he perceived
engagement in the middle press as more difficult. First, I expected that action-triggering time
would scale negatively with the current engagement rate in the middle press. I fit a GLM for
action-triggering time on all the non-repeated trials in which the monkey pressed on the
middle grip then on one side grip, and in which I could decompose RT into its subcomponents
(valid eye-tracking, monkey did look at both options and looked at chosen one during
execution). All the parameters found to have effects on action-triggering time (chosen and
unchosen Forces and Rewards, Side of press, action-selection time) as well as current
engagement rate for the middle press were included. Besides the same effects as what I
described before11, there was a negative effect of middle press engagement rate on action-
triggering time (β=-0.033; p=0.00012, Figure 52C), indicating that the monkey was faster at
triggering the chosen action if he had found engaging in the middle press less difficult and
would engage more often. Note that this effect remained true even if including middle press
RT and choice engagement rate to the model.
In order to confirm this result, I then looked for a relation between middle press RT
and action-triggering time, expecting a positive relationship. Again, I fit a GLM for action-
triggering time (same trials, sessions pooled) with chosen and unchosen Forces and Rewards,
Side of press, action-selection time, and middle press RT as parameters. Besides previously
described effects12, there was a positive effect of middle press RT on action-triggering time
(β=0.064; p<10-13, Figure 52D). This relationship remained true over and above engagement
rates in the middle press and in the choice.
11 Chosen Reward: β=-0.23; p<10-120; Chosen Force: β=0.031; p=0.00062; Unchosen Reward: β=0.049; p=0.00099; Unchosen Force: β=0.049; p<10-7; Side: β=0.095; p<10-6; Action-selection time: β=-0.78; p<10-150. 12 Chosen Reward: β=-0.24; p<10-128; Chosen Force: β=0.031; p=0.00040; Unchosen Reward: β=0.030; p=0.0015; Unchosen Force: β=0.048; p<10-7; Side: β=0.095; p<10-6; Action-selection time: β=-0.78; p<10-150.
Experimental results
193
Overall, these results indicate that the subjective difficulty to engage in the first step
of the task (middle press) was positively related to the subjective difficulty to engage in the
second step of the task (choice).
(2) Choosing between offers
In this task, in each trial, the monkey had to make a choice between two offers. I first
tested whether the monkey followed a speed-accuracy tradeoff when deciding and
questioned whether increased subjective difficulty to engage in the costly actions was
associated to changes in the speed-accuracy tradeoff, i.e. whether the current motivational
state of the monkey had effects on the resources he was willing to invest in the choices.
(a) Influence of Forces and Rewards on binary
choices between offers
First, I examined the effects of offer parameters (Forces and Rewards) on choices (left
or right option chosen), in order to be able to identify the value of offers based on choice
behavior. Overall, I expected the Forces to have a negative effect on choice probability and
the Rewards to have a positive effect on choice probability. Practically, I pooled the non-
repeated trials in which the cues appeared and the monkey made a choice (pressed on one
side grip) of all the sessions, and fit logistic regressions on choices (1: chose right side option,
0: chose left side)13. I compared three models: one with Rewards (left and right separately)
and Forces (left and right separately), one with the difference in Rewards (right–left) and the
difference in Forces (right-left), and one with the Reward minus the Force for both sides (left
and right, separately). I compared the fit of these regressions by comparing their BICs. The
model including all the offer parameters of both sides, separately, fit the data better
(BIC=3926.4) than the model using the differences between Forces (right-left) and Rewards
(right-left, BIC=3966.8), and than the model using the differences between Reward and Force
by side (BIC=4371.2).
In the best model, the right side Reward had a positive effect on right offer choice
probability (β=2.73; p<10-25), as well as the left side Force (β=1.15; p<10-12). Conversely, the
13 Identical results were obtained when including non-repeated trials too.
Experimental results
194
left side Reward (β=-3.05; p<10-29) and the right side Force (β=-1.04; p<10-10) had negative
effects on right offer choice probability (Figure 53A). Additionally, the constant, indicating a
potential side bias, was not significant (p=0.45). In other words, the probability to choose the
right side offer increased if it had a greater Reward and a smaller Force, i.e. a greater value,
and decreased if the left side option had a greater Reward and a smaller Force, i.e. a greater
value. Moreover, the monkey had no bias towards choosing one side above the other.
Figure 53: Binary choices between offers, and action-selection times.
(A) Coefficients for the logistic regression for choices (1 if the monkey chose the right side, 0
otherwise). The Reward offered on the right side had a positive influence on the probability to
choose the right side offer while the Force offered on the right side had a negative influence.
The opposite pattern was found for parameters of the left side offer, with a negative effect of
Experimental results
195
the Reward offered on the left side and a positive effect of the Force offered on the probability
to choose the right side option.
(B) Coefficients of the GLM for action-selection times. The chosen Reward and the unchosen
Force had negative effects, while the unchosen Reward and the chosen Force had positive
effects, indicating that a lower difference in value of the offers would lead to a longer action-
selection time.
(C) Coefficients of the GLM for action-selection times. There was a marginally significant
negative effect of the difference in the values of the chosen and the unchosen options (value
difference).
ctt: constant term ; F: Force; R: Reward; ***: p<0.001; **: p<0.01; *: <0.05 ; n.s.: non-
significant, error bars represent SEM.
These results would additionally indicate that, just like what I observed for choice
engagement, choices were more strongly influenced by Rewards than by Forces. In order to
properly assess that, I fit session-by-session logistic regressions for choice (1 if right side
chosen, 0 if left side chosen) on trials in which the monkey made a choice (pressed any side
grip). Using a two-way ANOVA, I compared the magnitude (absolute value) of the effects (β
values) depending on the side of the parameter (left or right) and the nature of the parameter
(Force or Reward). There was a significant effect of Force versus Reward, with stronger effects
associated to Rewards (F(1)=316.77; t-test on means of absolute values of β values, t(67)=-
2.3253; p<0.0001). There was no significant effect of side (F(1)=2.21; p=0.14). In other words,
Rewards had stronger effects on choices than Forces, and this pattern is reminiscent of what
I observed for choice engagement.
Overall, our results show that the monkey computed a tradeoff between Rewards and
Forces to choose between offers.
(b) Action-selection time had a tendency to increase
with choice difficulty
Second, I examined whether the pattern of action-selection times was compatible with
a speed-accuracy tradeoff. Humans and animals have been shown to increase the duration of
their decisions when difficulty (amount of information to process, conflict…) increases (Chittka
Experimental results
196
et al., 2009, 2003; Förster et al., 2003; Fudenberg et al., 2018). The notion of speed-accuracy
tradeoff stems from the postulate that accurate decisions require longer computations, and
are therefore more costly, at least in opportunities (Boureau et al., 2015). Practically, in this
task, during RTs, in most trials, gaze usually first alternated between the two cues, before
locking onto the chosen one. I posited that the period during which gaze was mobile between
cues corresponded to the action-selection period. Consequently, I looked at the relationship
of action selection duration with offer parameters and choice difficulty.
In order to investigate the relation between action-selection times and offered Forces
and Rewards, I pooled all the trials in which the monkey had looked at both options and had
looked at the chosen option during execution, from all the sessions in which I had valid eye-
tracking recordings (32/34 sessions). Additionally, I only selected trials in which the monkey
did not know about the offers displayed by the cues before they appeared, i.e. non-repeated
trials. Indeed, in repeated trials, the monkey probably did not need to select the chosen action,
as he may have done it before, and merely triggered it14. I fit a GLM on action-selection times
with chosen and unchosen Forces and Rewards as parameters, as well as the Side of press.
The effect of the chosen Reward (β=0.20; p<10-48) and of the unchosen Force (β=0.11; p<10-12)
were positive, and the effects of the unchosen Reward (β=-0.13; p<10-20) and of the chosen
Force (β=-0.09; p<10-9) were negative (Figure 53A). Additionally, action-selection time was
longer for right side choices (Side effect, β=-0.50; p<10-72). In other words, action-selection
was longer when the chosen option had a high value and when the unchosen option had a
low value. Thus, I compared the fit of the model presented above and that of a model including
only the difference in offer value (determined using choices, see methods) or the absolute
value of the difference between the values of the chosen and unchosen options and the side
of the chosen offer. I found that the model including all the parameters separately fit the data
better (Model with all the parameters: BIC=13 254; model with value difference: BIC=13 528,
model with absolute difference of values: BIC=13 529).
Additionally, to be able to compare the magnitude of the effects of Forces and Rewards
on action selection time, I fit a GLM per session for action selection time with chosen and
14 This hypothesis was confirmed by fitting a GLM for action selection times in repeated trials, with chosen and unchosen Rewards and Forces, and chosen side as parameters. No parameter had a significant effect (p>>0.05). This provides additional evidence that monkey selected his course of actions during the period of gaze position alternation between offers.
Experimental results
197
unchosen Forces and Rewards. I selected trials in which the monkey saw the cues, made a
choice, had looked at both options, and was looking at the chosen cue when triggering the
action. I then performed a two-way ANOVA (analysis of variance test) on the GLM coefficients
with chosen versus unchosen parameters and nature of parameters (Force or Reward) as
factors. The ANOVA failed to detect any effect of the “Force versus Reward” parameter
(F(1)=1.06; p=0.31) or of the “chosen versus unchosen” parameter (F(1)=2.77, p=0.09). In
other words, contrary to what was observed for choices between offers for which Rewards
had a stronger effect, action-selection duration was affected by Rewards and Forces with
virtually equivalent magnitudes.
Although the model explaining action-selection time with the value difference
between the chosen and unchosen option was not the model fitting the data best, there was
a marginally significant positive effect of the value difference on action-selection time
(β=-0.024; p=0.082, Figure 53B). This provides some indication that the monkey had a
tendency to be faster when the difference in value between the options was higher. These
results indicate that the monkey took more time to choose between options when the choice
was more difficult, with closer values of the two options.
(c) Action-selection time did not influence choice
accuracy
Next, I looked for evidence that increased time spent on the choice led to more
accurate decisions, which could have indicated that longer action-selection times were
associated with increased exertion of cognitive control for the decision. Using the subjective
values of the offers inferred from choices (see method in II.B.2.d), I determined whether the
monkey had chosen according to the average value that he had attributed to each offer, i.e.
whether he had made an accurate choice. First, I fit a logistic regression for choice accuracy (1
if option with highest subjective value was chosen, 0 otherwise) with only action-selection
time as parameter, on all the non-repeated trials in which the monkey made a choice and in
which action duration time could be determined. Action-selection time had a negative effect
on choice accuracy (β=-0.19; p<10-6), indicating that the longer the monkey took to reach a
choice, the less accurate his choice would be (Figure 54A).
Experimental results
198
In order to remove any difficulty-related effects, I then added the absolute value
difference (value of best option value minus value of the worst option, ΔVbest-worst) and offer
parameters (Forces and Rewards of best and worst options). The effect of action-selection
time was no longer significant (β=-0.07; p=0.13, Figure 54B). ΔVbest-worst had a positive effect
on choice accuracy (β=4.65; p<10-4). Regarding individual offer parameters, the Reward of the
best option (β=2.30; p<10-64) and the Force of the worst option (β=1.05; p<10-57) had positive
effects, while the Force of the best option (β=-0.97; p<10-55) and the Reward of the worst
option (β=-1.95; p<10-63) had negative effects. These results indicate that action selection time
had no effects on choice accuracy over and above effects of choice difficulty, i.e. the
relationship between action-selection time and accuracy was due to the fact that action-
selection time was longer in case of difficult choices (see Figure 53B&C), associated to lower
choice accuracy.
(d) Absence of relationship between engagement
difficulty and choice accuracy
Earlier, I established that difficulty, i.e. required effort, to engage, probably reflected a
global motivational state with effects on all engagement-related processes throughout the
trial. Next, I questioned whether this motivational state had effects on the cognitive effort
invested in choices between offers. Specifically, I would have expected that periods of time
during which the monkey found it more difficult to engage were associated to lesser choice
accuracy. Therefore, I tested whether choice accuracy varied with current middle press and
choice engagement rates, middle press RT, and action-triggering time (side press). I fit a
logistic regression for choice accuracy as described above, and added either one of these four
parameters to the regression. In none of the cases did I find a significant relationship between
choice accuracy and such proxies for the effort required to engage (p>0.55 in all cases). This
result indicates that the motivational state that influences the engagement of the monkey and
more specifically the subjective difficulty of engaging was probably not related to the
mobilization of resources for the choices.
Experimental results
199
(e) Increased action-selection time is associated to
decreased sensitivity of choices to Forces
One major assumption of the previous analyses was that choices were considered to
be accurate, and therefore possibly associated to more control exertion, if they were
consistent with the mean value of the offers. Cognitive control is required to inhibit decision
biases, and avoid default (automatic) courses of action, in order to promote strategies
compatible with the goals, usually reward rate maximization (part I.A.2.d)(1) for review). In
that task, the monkey perceived the high-Force conditions as aversive, thus making his choices
by avoiding high-Force options. Moreover, there was a tendency for combinations of offers
with a high sum of Forces to be refused more often. In that sense, the default course of action
would be choosing the low-force options. Thus, it is possible that increasing cognitive control
investment for decisions would have facilitated high-Force choices. If that were true,
considering the extensive literature on the relationship between the time spent on decisions
and the cognitive control allocated to the decision process (Hunt et al., 2012; Krajbich et al.,
2010; Krajbich and Rangel, 2011; Milosavljevic et al., 2010; Morris et al., 2019; Philiastides and
Ratcliff, 2013), I posited that increased action-selection time should be associated with
decreased effects of Forces on the probability to choose the high-Force option (Force
sensitivity), but should not affect Reward sensitivity.
To test this hypothesis, I categorized the choices made by the monkeys in two ways:
high- versus low-Force choices (if he chose the option associated to the highest Force or not)
and high- versus low-Reward choices (if he chose the option associated to the highest Reward
or not). I pooled all the non-repeated trials in which the monkey made a choice (pressed on
one side grip after cues onset). First, I split trials depending on whether the high-Force option
had been chosen or not (excluding trials in which the Forces offered on both sides were
identical). I then fit a logistic regression for high-Force choices (1 if the high Force option was
chosen, 0 otherwise) with the difference between the high and low Forces and difference of
Rewards of the high-Force and low-Force options. Additionally, based on our hypothesis, I put
the interactions between action-selection time and the differences in Forces and in Rewards
of the offers.
Experimental results
200
Figure 54: Relationship between action-triggering time and choice accuracy: tenuous evidence
for a speed-accuracy tradeoff, and evidence for a “speed-cognitive control production”
tradeoff.
(A,B) Logistic regressions for choice accuracy (1 if the monkey chose the option with the highest
value, 0 otherwise). (A) When action-selection time was alone in the model, there was a
negative relation between action-selection time and choice accuracy. (B) This relationship did
not stand over and above choice difficulty (Difference in value between best and worst option,
Delta V best-worst) and offer parameters (Forces and Rewards).
Experimental results
201
(C) Logistic regression for high-Force choices (1 if the monkey chose the option with the highest
Force offered, 0 otherwise). Over and above effects of the difference in Forces (F difference,
high – low force) and in Rewards (R difference, Reward of high force option – Reward of low-
Force option), there was negative effect of the interaction between action-selection time and
the difference in Forces.
(D) Interaction plot showing the effects of the interaction between action-selection time (AST)
and the Force of the high-Force option (y-axis) on the probability to choose the high-Force
option. Action-selection times were split between those above (dotted line) or below (full line)
the median (m). The effect of the difference of Force was smaller when action-selection time
was longer.
ctt : constant term ; F: Force; R: Reward; ***: p<0.001; error bars represent SEM.
Expectably, there was a positive effect of the difference in Rewards (β=-0.61; p<10-13)
and a negative effect of the difference in Forces (β=4.26; p<10-124, Figure 54C). This indicates
that the monkey was more willing to choose the high-Force option if it was associated to a
higher Reward, and was and less willing to choose it if the alternative was associated to a
higher Reward. Moreover, the monkey was more willing to choose the high-Force offer if the
highest Force proposed was close to the lowest force proposed. Additionally, there was a
negative effect of the interaction between the difference in Forces and action-selection time
(β=-0.16; p=0.041). Specifically, longer action selection-times were associated to decreased
effects of the difference in Forces (Figure 54D). There was no effect of the interaction between
action-selection times and Rewards (β=-0.085; p=0.56). Note also that there was no main
effect of action-selection time if it was added to the model, indicating that longer action-
selection times did not merely bias the choice process towards high-choice options, but
decreased the sensitivity to Forces, possibly because longer action-selection times were
associated to increased cognitive control production.
I repeated the procedure, this time splitting trials between ones in which the highest
offered Reward was chosen (high-Reward option chosen) and ones in which the lowest
offered Reward was chosen (low-Reward option). I put the difference in Rewards (biggest –
smallest Reward) and the difference in Forces (Force of the high-Reward option – Force of the
low-Reward option) as well as the interactions between action-selection time and Force and
Experimental results
202
Reward difference as parameters. There was a negative effect of the difference in Forces (β=-
1.88; p<10-38) and a positive effect of the difference in Rewards (β=1.52; p<10-24) but no effects
of their interactions with action-selection time (respectively: β=0.13; p=0.24 and β=-0.002;
p=0.99).
These results indicate that increased time spent on the decisions was specifically
associated with decreased sensitivity of the choices to the Force parameter of the offers,
making the monkey more likely to engage in courses of action requiring high physical effort
when he had taken longer to selection a course of action. Considering that higher cognitive
control is associated weaker biases towards responding according to the default (reviewed in
part I.A.2.d)(1)), and since the default response would be avoiding high-Force offers, this result
would indicate that longer decision times were associated to increased control exertion and
lesser biases to avoid high-Force options.
(3) Execution of costly actions
In this task, two costly actions had to be completed: a press on the middle grip, with a
minimal amount of force, and a press on one of the side grips, with a force above the threshold
requested by the cue for at least 80ms. I tested whether I could find evidence of a relationship
between the subjective difficulty of engaging in actions and the physical effort produced for
the execution of the actions. Specifically, I expected that when the monkey perceived
engagement and action-triggering as more difficult, he would also perceive action execution
as more difficult and would exert less force and have lower performance.
(a) Middle press exerted force was negatively
related to difficulty to engage in middle press
First of all, I examined the relationship between the force exerted on the middle grip
and the subjective difficulty of engaging in the middle press, approximated by engagement
rates in middle presses and middle press RTs. I fit GLMs for the maximum force exerted on the
middle grip with either only the block-wise engagement rate in the middle press or only the
middle press RT. There was a positive effect of middle press current engagement rate (β=0.08;
p<10-12), or a negative effect of middle press RT (β=-0.12; p<10-25) on the maximum force
exerted on the middle grip (Figure 55A&B respectively). This indicates that faster RTs, i.e.
Experimental results
203
lower subjective difficulty to engage, was also associated to higher force exertion. In other
words, the monkey exerted less force on the middle grip if he perceived engaging in the middle
grip press as more difficult.
Figure 55: Negative relationship between maximum exerted force and engagement difficulty.
(A,B) GLMs for maximum exerted force on middle grip. The monkey exerted more force on the
middle grip if (A) current engagement rate was higher and (B) RT to the middle press had been
shorter, suggesting that increased difficulty to trigger the middle press led to decreased
exerted force.
(C) GLM for maximum exerted force on side grip. Action-triggering time had a negative effect
on the force exerted on the side grip, suggesting that increased difficulty to trigger the side
press led to decreased force exertion.
Experimental results
204
(D) Logistic for performance of side press. Action-triggering time had a negative effect on
performance over and above effects of offer parameters and side.
(b) Side press exerted force was negatively related
to difficulty to engage in side press
Second, I examined whether increased subjective difficulty to engage in the chosen
action was also associated to decreased exerted force for the chosen action (side grip). I fit a
GLM for the maximum exerted force on the side grip at the moment of the execution of the
chosen action, with offer parameters (chosen and unchosen Rewards and Forces), side of
press, and action-triggering time. I used all the trials15 in which I could identify action-
triggering time, and in which the monkey had pressed on the side grip.
There was a negative relation between action-triggering time and the maximum force
exerted on the side grip (β=-0.040; p<10-5, Figure 55C). Additionally, the chosen Reward had
a positive effect (β=0.042; p<10-4), as well as the chosen Force (β=0.70; p<10-150), the
unchosen Force (β=0.049; p<10-7) and side of press (β=0.44; p<10-123). The unchosen Reward
had a negative effect (β=-0.057; p<10-8).
Note that middle press RTs, middle press engagement rate and choice engagement
rate were not correlated to side press exerted force over and above the positive relationship
between middle press exerted force and side press exerted force.
(c) Performance was negatively related to action-
triggering time
Over all sessions, the monkey’s success rate was 99.81% chosen Force 1, 94.8% for
chosen Force 2 and 71.7% for chosen Force 3. Based on the effect of action-triggering time on
exerted force, I could expect either one of two different effects on performance: no effect if
the exerted force decreased but was maintained above threshold, i.e. if force was optimized
to be minimized without changing performance, or a decrease in performance if the exerted
force was decreased regardless of the instruction. To examine this, I pooled all the trials in
15 Identical results were obtained using only non-repeated trials.
Experimental results
205
which the monkey made a side press, with a valid eye-tracking signal and in which decision
time was determinable, and fit a logistic regression for performance, with chosen Forces and
Rewards, side of press, and action-triggering time as parameters. There was a negative effect
of action-triggering time on performance (β=-0.23; p<10-5, Figure 55D), as well positive effects
of the chosen Reward and of the unchosen Force (β=0.13; p=0.035 and β=0.22; p<10-4,
respectively), and negative effects of the chosen Force and the unchosen Reward
(respectively: β=-1.80; p<10-92; and β=-0.14; p=0.026). The Side of the press had no significant
effect (β=0.15; p=0.16). This indicates that increased time taken to trigger the action was
associated to decreased performance, meaning that the observed decrease in exerted force
did not correspond to an optimization of resource mobilization for the action, but rather an
overall decrease in effort production.
Overall, the results described above indicate that increased subjective difficulty of
engaging in actions and triggering them led to decreased force exertion during actions, and
decreased performance, when the force was instrumental for the action. This suggests a
decrease in physical effort production for action execution when engaging in the action was
more costly. This could suggest either that engaging when subjective difficulty was high was
overall associated to a low motivational state, and that this state also affected physical effort
production. Alternatively, it could indicate that making the effort to engage when subjective
difficulty was high depleted a certain resource critical for physical effort, resulting in a lower
capacity to exert force.
(4) Summary of behavioral results
To summarize, in this task, three categories of processes potentially required effort
exertion. First, engaging in costly actions probably required cognitive control exertion, as the
default (automatic) response would have been avoiding the high-physical effort options (see
part I.A.3 for relationship between engaging in actions requiring physical effort and cognitive
control). I showed that the subjective difficulty of engaging could be approximated on a trial-
by-trial basis by the time taken to trigger each action (middle press RT and action-triggering
time for side press). Indeed, the monkey took longer to trigger both actions in periods of
overall low willingness to engage, i.e. high subjective difficulty of engaging. The subjective
Experimental results
206
difficulty to engage in both steps of the task was strongly correlated, suggesting that the
subjective difficulty of engaging in each step depended on a global effort variable.
Second, making choices between offers probably required cognitive control (see part
I.A.2.d)(1) for the role of cognitive control on optimizing decisions). I showed that the monkey
maybe followed a speed-accuracy tradeoff, as the time spent deciding increased for difficult
choices but choice accuracy still decreased, suggesting that the time spent deciding was not
increased enough to cope with the increase in difficulty. Moreover, I could not find any
consistent relation between the subjective difficulty of engaging in actions and the accuracy
of the choices of the monkey. Conversely, increased time spent on choices led to decreased
Force sensitivity, but did not alter Reward sensitivity. In that sense, action-selection time may
reflect the cognitive control invested in the decision to avoid choosing default courses of
actions associated to low physical effort.
Third, executing the forces probably represented a physical effort, at least for the side
press, since monkeys discounted offered Rewards with required Forces in their choices. I
showed that increased subjective difficulty of engaging in actions led to decreased force
exerted on the grips. This indicates either that states of lower motivation were associated with
lower physical effort exertion or that exerting cognitive control to trigger the chosen action
led to decreased capacity for physical effort (see part I.A.3 for negative effects of cognitive
control exertion on physical effort).
b) Neurophysiology of the LC
54 LC single-units were recorded in the two-alternative choice task. LC neurons were
identified prior to task onset based on anatomical criteria (located just ventrally to the motor
part of the mesencephalic trigeminal nucleus which is sensitive to face and mouth
movements) and well-described electrophysiological criteria (low baseline rate <5Hz, large
waveforms and activation-inhibition responses to sudden sounds or other stimuli, Bouret and
Richmond, 2009; Grant et al., 1988). The position of the electrode in the LC was confirmed in
several sessions using clonidine injections (Bouret and Richmond, 2009; De Sarro et al., 1987;
Gatti et al., 1988; Varazzani et al., 2015), which caused a strong decrease of LC neurons’ firing
rate (Figure 47B). Moreover, if the monkey became lethargic or fell asleep, the firing of the
recorded LC neurons strongly decreased. The average firing rate of all 75 neurons was
Experimental results
207
4.78±0.73 spikes/s over the course of recording sessions, and the baseline firing rate of the
neurons remained quite constant when the monkey was not drowsy. I never observed any
change in firing mode (e.g. tonic vs phasic), but LC neurons tended to exhibit a decrease in
rate at the beginning of the task, compared to pre-task rate.
Throughout analyses of LC activity in that task, I first looked for evidence of an
encoding of the effort produced to engage in actions and trigger them, choose between offers,
and execute them, or at least of the subjective difficulty of these processes. I expected LC
neurons to display a domain-general encoding of effort production throughout the task.
Regarding cognitive control exertion to engage in actions and trigger them, prior to middle
press onset and prior to the chosen action, I expected LC neurons to encode the instantaneous
subjective difficulty of triggering the middle press and the chosen action, respectively.
Specifically, I expected LC neurons to encode middle press RT before the middle press, and
action-triggering time before the chosen action. Importantly, I did not expect LC neurons to
encode the current engagement rates in the middle press and the choice, respectively. Indeed,
despite their correlations with middle press RT and action-triggering time, these variables
represented long-term motivational states rather than instantaneous subjective difficulty. In
other words, I expected LC neurons to encode the cost of ongoing processes with a fine time
scale. Regarding physical effort production to execute the presses, I expected LC neurons to
encode the exerted force on the grip during force exertion.
Second, I expected LC activity to have delayed effects on effort production. In other
words, I expected increased LC activity early in the trials (e.g. around the middle press) to have
a facilitative effect on effort production later in the trial (e.g. engaging in the choice and
choosing between options), regardless of the challenge at hand at that moment. I
systematically examined the relation between LC activity at each step of the task (middle press
onset and execution, cues onset, chosen action onset) and effort-related behavior later on
(engaging, choosing between offers, triggering and executing the chosen action).
Third, I hypothesized that the signal regarding effort production constituted by LC
activity around chosen action execution could serve as a feedback to the effort allocating
system. Practically, this means that I expected the strength of the activation of LC neurons
around chosen action execution (over and above task parameters and the exerted force) to
predict the effort sensitivity of the binary choice made in the next trial. Additionally, in order
Experimental results
208
to correctly adjust effort allocation, the system would need to receive feedback on errors.
Since commitment error signaling by LC neurons had already been reported in the past (Jahn
et al., 2020), in order to confirm and extend this finding I looked for commitment and
execution error signaling by LC neurons in my paradigm.
(1) Middle press-related activity: encoding of middle press
RT and enhancement of engagement-related processes and of
Force exertion
In the previous part, I showed that the subjective difficulty of triggering the middle
press varied across sessions and could be approximated on a trial-by-trial basis using middle
press RT. Given the results of the previous study (Bornert and Bouret, in prep) regarding the
positive relationship between LC activity before the onset of a costly action and RT, I expected
LC neurons to encode middle press RT before middle press onset. Additionally, I looked for
short-term (force exerted on the middle grip, immediately after triggering the press) and
longer-term (choice engagement, choice accuracy, chosen force exertion) effects of such LC
activity on behavior, given our hypothesis that LC activity energizes behavior, and specifically
facilitates effort production after the activation.
(a) LC neurons increased their rate before and
during middle press
First of all, I assessed rate changes from baseline around the middle press. To quantify
the proportion of neurons that displayed evoked responses around the press, and test for a
population-wise modulation of rate, I computed the neuron-by-neuron spike count per trial
in a pre-press ([-300;0ms] from press onset) and a post-press ([0;300ms] from press onset)
epoch, and in reference windows (baseline, [-300;0ms] from middle red dot onset). I then
compared the spike count in the baseline window and the spike count in the two windows of
interest with a paired t-test. At the level of the population, coherence of evoked responses
was tested by performing a t-test on the neuron-by-neuron t values previously obtained, and
comparing their mean against zero.
Before press onset, rate changed from baseline for 27/54 neurons (20 increased and 7
decreased), and there was a coherent increase in rate at the level of the population (second
order: t(53)=2.56; p=0.0132, Figure 56A). After press onset, rate changed from baseline for
Experimental results
209
36/54 neurons (30 increased and 6 decreased), and there was a coherent increase in rate at
the level of the population (second order: t(53)=7.71; p<10-9, Figure 56B).
During the press, I questioned whether LC activity was better aligned onto the onset
of the press or the attainment of the maximum exerted force on the grip. In that aim, in all
the non-repeated trials in which the monkey made a side press, for all 54 neurons, I identified
the time of the attainment of the maximum in exerted force on the middle grip, and for each
spike in the press-onset to maximum of force exerted epoch, I computed the absolute value
of the time elapsed from the spike to the onset of the press and to the peak of the press.
Neuron-by-neuron, I then calculated the density of the distribution of these times to either
event, and identified the maximum of each density function (method adapted from Bouret et
al., 2012). Across neurons, I then performed a t-test on the maxima of the density functions.
These density functions included the same number of data points, since I used the same spikes
only aligned on different events. Therefore, a higher maximum of the density function would
signal a narrower distribution, and therefore a better alignment of the timing of spikes. Across
the population, spikes were better aligned onto the peak of the press than its onset (t(53)=-
2.64; p=0.011).
(b) LC neurons encoded middle press RT before
middle press onset
LC neurons activated just prior to middle press onset, which, in non-repeated trials,
signaled engagement in trials independently of value evaluation. I reasoned that the
frequency of engagement in middle presses across trials as well as the middle press RT (time
elapsed between Go signal - central green dot - and initiation of middle press) should
represent the cognitive control invested in triggering the middle press. Under the hypothesis
that LC neurons signal resource mobilization for costly actions, or at least the difficulty of
triggering costly actions, I questioned whether LC neurons encoded engagement rate in
middle presses, and middle press RT, just before middle press onset. Although I showed earlier
that these two measures were correlated negatively, their encoding by LC neurons would
reflect slightly different intuitions: while current engagement rate (computed over several
trials), represents a motivational state impacting the subjective difficulty of triggering actions,
trial-by-trial RT may represent instantaneous difficulty or resource mobilization for triggering
action initiating the trial.
Experimental results
210
Figure 56: Activity of LC neurons around middle press onset: LC neurons activated around the
middle press and encoded middle press RT before the middle press.
(A) Sliding window representation of the mean rate across time of all the LC units recorded (y-
axis, 1 neuron per line), around middle press onset. Rate is z-scored by neuron. The green dot
indicates the mean timing of the onset of the green dot (based on the mean middle press RT
of the session).
Pre action onset
Post action onset
Experimental results
211
(B) Density plots (bandwidth=5) of t values of the comparison of rate in the pre (yellow) and
post (dark red) middle action onset epochs, with baseline (pre red dot onset) rate. Second order
t-test on rate indicated that, at the level of the population, around middle press onset, rate
significantly increased.
(C) GLMs for pre-middle action spike count: mean of individual beta values and SEM across
neurons. There was no significant effect of block-wise engagement frequency at the level of
the population.
(D) Raster plot of example LC unit displaying a positive encoding of RT before middle press
onset. More spikes (black dots) were present before middle press onset (pink line) after long
RTs (longer time during green dots representing go-signal for middle press and pink line
representing middle press onset).
(E) GLMs for spike count: mean of individual beta values and SEM across neurons. LC neurons
encoded RT positively (significant effect across the population) before middle action onset.
*: p<0.05; **: p<0.01; n.s.: non significant, error bars represent SEM.
First, I cut testing sessions in halves and computed the frequency of engagement in the
middle press in non-repeated trials in each block i.e. the proportion of non-repeated trials in
which the monkey made a middle press. I then fit, for each neuron, a GLM for spike count in
a pre-middle action epoch ([-200;0ms] epoch before middle action onset, based on raster
examination) with block-wise frequency of engagement in middle presses as parameter. I had
to exclude 3 neurons recorded in sessions with insufficient differences in engagement rate
between blocks. For 13/51 neurons, block-wise engagement frequency had a significant effect
(7 positive, 6 negative). At the level of the population, there was a no significant effect (second
order t-test on T values: p=0.30, Figure 56C).
Second, I fit a neuron-by-neuron GLM for spike count in the same pre-middle action
epoch, with only RT for the middle press as parameter. I only included trials for which the
monkey did not know about the future offers (cues) i.e. that had not been repeated, and in
which he pressed on the middle grip during the correct period (during go-signal presentation).
9/54 LC neurons encoded RT positively in that epoch (8 positively, 1 negatively). The activity
of an example unit is showed in Figure 56D. The positive encoding of RT was consistent at the
level of the population (second order t-test on T values: t(53)=2.75; p=0.0081, Figure 56E).
Experimental results
212
Note that this encoding was not present earlier before press onset or before red dot onset,
nor was an encoding of current willingness to engage in the middle press.
Overall, these results reproduce the result obtained in the first study (Bornert and
Bouret, in prep). They point towards an encoding of instantaneous subjective difficulty rather
than current (longer-term) motivational state, before the monkey engages in the middle
press.
One of the hypotheses of this work concerns the fact that LC neurons encode
instantaneous effort production. If the monkey produced an effort to execute the middle
press, the force exerted at that time should represent the physical effort associated to
pressing the middle grip. I therefore examined whether LC neurons encoded the maximum
force exerted on the middle grip, during force exertion. I fit a neuron-by-neuron GLM for spike
count in a [0;300ms]16 epoch after press onset (during force exertion) with only the maximum
exerted force on the middle grip as parameter, on non-repeated trials in which the monkey
engaged in the middle press during the dedicated time interval. The effect of the maximum
exerted force on the middle grip was significant for 8 neurons (2 negative, 6 positive) and was
not consistent at the population level (second order: p=0.32, Figure 56E). Note that, during
this epoch, RT was no longer encoded, and the exerted force on the middle grip was not
consistently encoded prior to action onset either (p>>0.05 at the population level). Similarly
to what we discussed in Study 1 (Bornert and Bouret, in prep), the minimal force that had to
be applied on the middle grip may not have required effort production.
(c) Middle press activity and engagement throughout
the trial: higher pre-middle press activity was associated
to higher probability to engagement in the choice
Prior to middle press onset, I showed that LC neurons encoded RT, which I hypothesize
to represent the effort made, or at least required, to trigger the action, and therefore engage
in a trial. Next, I questioned if this activity around the beginning of trials could influence
behavior later on during trials. Specifically, I hypothesized that increased LC activity would
16 Identical results were obtained when using a time window selectively including the period before the peak in force (see Figure 48A for force signals across time), around 200ms from action onset (spike count in a [0;200ms] epoch) or selectively the period after the peak in force ([200;400ms] after action onset).
Experimental results
213
lead to increased effort production later on, possibly a few seconds later. In the particular case
of our task, I hypothesized that pre-middle press activity, putatively reflecting the effort
invested in starting the trial, would have beneficial effects on the next processes requiring
effort production, regardless of the nature of the challenge (mental or physical): force exertion
on the middle grip and choice engagement. Since I questioned whether LC neurons as a
population could have effects on behavior, the general method I used for testing such effects
was pooling the relevant trials from all the sessions, i.e. from all the neurons, before fitting
regressions on the behavioral variable of interest, with spike count in windows of interest as
one of the parameters. Based on results from the previous study regarding the positive
correlations between the strength of encoding of task parameters and of behavioral variables
and baseline as well as evoked firing rate (see part II.A.2.c)(3)), I expected more active neurons
to have stronger effects on behavior, and therefore did not normalize spike count per neuron
before pooling the trials.
I looked for a link between pre-middle press onset activity and choice engagement,
over and above effects of the overall difficulty to engage and of offer parameters. I pooled
trials from all the sessions and fit a logistic regression for choice engagement with the sum of
Forces and the sum of Rewards, current engagement rate in middle press or middle press RT,
and pre-middle press onset spike count ([-200;0ms] epoch, in which RT is encoded by LC
neurons) as parameters. Only non-repeated trials in which the monkey pressed on the middle
grip were included.
Over and above the effects of the sum of Rewards (β=0.71; p<10-13), of Forces (β=-0.09;
p=0.25), and of the current rate of engagement in the middle press (β=0.13; p=0.07), there
was a positive effect of spike count on choice engagement (β=0.20; p=0.04, Figure 57A). The
same was true when including middle press RT in the model as a proxy for the difficulty to
engage (sum of Rewards: β=0.71; p<10-13, sum of Forces: β=-0.08; p=0.26, middle press RT:
β=-0.24; p=0.00014, spike count: β=0.20; p=0.04, Figure 57B). This indicates that increased
spiking before middle press onset was associated to a higher probability of engagement in the
choice. Importantly, these effects were not present in baseline activity (pre-middle red dot
activity) or in post-middle press onset activity (during middle press).
Experimental results
214
Figure 57: Increased LC activity before middle press onset was associated to decreased
difficulty to engage in the choice, as choice engagement increased and action-triggering time
decreased.
(A,B) Logistic regressions for choice engagement, on all non-repeated trials in which the
monkey pressed on the middle grip. Spike (spk) count in the [-200;0ms] epoch before middle
press onset had a positive effect on choice engagement, over and above the positive effect of
the sum of Rewards (Sum R) and the negative effect of middle press RT, and (A) the current
engagement rate in the middle press or (B) middle press RT.
(C) GLM for action-triggering time. Spike count in the [-200; 0ms] epoch had a negative effect
on action-triggering times, over and above effects of task parameters, chosen side and over
Experimental results
215
and above the strong negative correlation between action-selection and action-triggering
times. This indicates that increased LC activity led to faster responses.
*: p<0.05; **: p<0.01; ***: p<0.001, n.s.: non significant, Spk: spike, F: Force, R: Reward
Such effects of LC activity around the middle press could indicate either that engaging
in the choice and therefore triggering the chosen action was perceived as less difficult, or that
the monkey engaged more effort despite equal perceived difficulty when the LC neurons had
spiked more. We therefore fit a GLM for action-triggering time with pre-middle press activity
as parameter, as well as chosen and unchosen Forces and Rewards, action-selection time, and
side of press as parameters. Besides identical effects of offer parameters, action-selection
time and side as described earlier17, spike count before middle press onset had a negative
effect on action-triggering time (β=-0.022; p=0.011, Figure 57C), indicating that increased
spike count before middle press onset led to decreased time taken to trigger the chosen
action. In other words, after increased spiking of LC neurons around the middle press, the
monkey was both more willing to engage in the choice and faster to trigger the chosen action,
confirming the relation between middle press LC activity and subjective difficulty to engage in
the choice.
This effect was also present in the post-middle press onset epoch ([0;200ms] from
middle press onset, β=-0.041; p<10-5), indicating that spike count in an epoch around middle
press onset ([-200;200ms] from middle press onset) had a negative influence on action-
triggering time. I confirmed that by fitting a GLM for action-triggering time with spike count in
the [-200;200ms] epoch around middle press onset, chosen and unchosen Forces and
Rewards, action-selection time and side of press as parameters. The effect of spike count was
significantly negative (β=-0.036; p=0.0004).
Note that there was also a negative effect of baseline spike count on action-triggering
time, but that effect did not remain significant if middle press spike count ([-200;200ms]
epoch) was included in the model. Additionally, the fit of the GLM with baseline spike count
was not as good as the fit of the model with peri-middle press spike count (BICbaseline=8981.8;
BICmiddle press=8974.1).
17 Action-selection time: β=-0.78; p<10-150; Chosen Reward: β=-0.23; p<10-120; Chosen Force: β=0.030; p=0.00089; Unchosen Reward: β=-0.032; p=0.0010; Unchosen Force: β=0.049; p<10-7; Side: β=0.094; p<10-6.
Experimental results
216
(d) Middle press activity and force exertion: higher
pre-middle press activity was associated to increased
middle press exerted force
Next, I examined the relationship between pre-middle press LC activity and exerted
force throughout the trial. First, I tested the relationship between activity before middle press
onset and middle press exerted force, over and above the negative correlation of middle press
exerted force with middle press RT. I pooled all the non-repeated trials in which the monkey
pressed on the middle grip and fit a GLM for the maximum exerted force on the middle grip
with middle press RT and pre-middle press spike count ([-200;0ms] from middle press onset)
as parameters. Over and above the previously described negative effect of middle press RT
(β=-0.12; p<10-17), there was a positive effect of spike count on middle press exerted force
(β=0.084; p<10-8, Figure 58A).
Figure 58: Pre-middle press spike count had a positive relation with middle press exerted force.
(A,B) GLMs for middle press exerted force. There was a positive effect of pre-middle press spike
count ([-200;0ms] from middle press onset) on middle press exerted force, (B) as well as a
positive effect of the spike count during the press. Baseline rate had no effects.
*: p<0.05; ***: p<0.001, n.s.: non significant, Spk: spike, F: Force, R: Reward
This positive relationship between the middle press exerted force and spike count was
still present during the press, and specifically during the period of the press that preceded the
Experimental results
217
attainment of the peak of force ([0;200ms] after press onset (see Figure 48A for time course
of force exertion), over and above pre-middle press spike count (spike count before action:
β=0.039; p=0.029; spike count during action: β=-0.072; p<10-4; middle press RT: β=-0.12;
p<10-16). Additionally, adding baseline spike count (pre-middle red dot spike, [-300;0ms]
epoch) did not change the effect of pre- and post-middle action onset spike count on exerted
force (spike count before action: β=0.040; p=0.040; spike count during action: β=-0.073;
p<10-4; middle press RT: β=-0.12; p<10-16, Figure 58B). Moreover, the effect of baseline rate
on middle press exerted force was not significant (β=-0.003; p=0.87). These results tend to
indicate that peri-middle press activity was associated to higher middle press exerted force,
when regressing out effects of middle press RT, and considering the neurons as a population.
Overall, across the population, the spike count of LC neurons before middle press onset
scaled positively with middle press RT, a proxy for the instantaneous difficulty to trigger the
middle press. Increased activity in that epoch was associated to increased probability to
engage in the choice regardless of offer parameters. Moreover, increased spike count was also
associated with increased exerted force for the middle press.
(2) Activity around cues onset
Given the indirect (pharmacological or pupilometry studies, see part I.B.5.a)) evidence
for a role of LC activity in action-selection, and cognitive control exertion for choices against
biases, I looked for direct evidence for such roles in the activity of LC neurons and its relation
to behavior. The monkey only knew about the parameters (Forces and Rewards) of the offers
after cues had appeared (in non-repeated trials), and had putatively reached a decision (made
a choice) by the time he locked his gaze onto the chosen option (choice onset). Therefore, I
focused on the activity of LC neurons between cues onset and choice onset, in non-repeated
trials. I first examined the pattern of evoked responses during that period, then analyzed the
modulation of activity by offer parameters, and finally tested for effects of this activity on
choices, and especially on choice accuracy and Force sensitivity.
(a) Responses to cues onset and choice onset
Examination of LC activity during the second step of the task revealed that a certain
proportion of LC neurons displayed bursts of activity (a few spikes) between cue onset and
choice onset (last saccade onto chosen option, i.e. end of action-selection time).
Experimental results
218
I examined whether I could identify if the activity of LC units between cues onset and
choice onset was better aligned onto cues onset or onto choice onset. In that aim, using the
same method as the one used for middle press related activity, I compared the distribution of
the timing of spikes when aligned on cues onset or on choice (last saccade onto chosen option)
onset. I could not find any difference between the distribution of the event-to-spike times for
the activity comprised between those two events, i.e. I could not conclude on a better
alignment of spikes onto cue onset or choice onset (t(49)=1.23; p=0.22).
I then quantified the proportion of neurons that changed their activity after cues onset.
I used the same procedure as the one described for middle press onset responses. In brief,
neuron-by-neuron, I compared rate in windows of interest to rate in a baseline window, using
paired t-tests. Population rate changes were assessed by performing a second-order t-test on
the T values previously obtained. Note that the windows were assigned based on raster
examination. Around cue onset, I quantified activity in a [0;500ms] post-cues onset epoch,
and compared it to a [-500;0ms] epoch preceding cues onset18 (Figure 59A). 29/54 neurons
exhibited a change in rate (rate decreased for 9 neurons and increased for 20 neurons) and
this change was marginally significantly positive at the level of the population (second-order
t-test: t(53)=1.96; p=0.055, Figure 59B).
(A) (B)
Figure 59: Rate of LC neurons around cues onset, and results of first and second order t-tests
assessing rate changes: after cues onset, LC activity increased.
18 This baseline epoch was entirely comprised during the presentation of the middle blue dot, and started at least 400m after the middle press.
Experimental results
219
(A) Activity of the 54 recorded LC units around cues onset. Rate is z-scored by neuron. The
magenta dot indicates the mean RT of the session, i.e. the mean timing of action onset.
(B) Density plots (bandwidth=5) of the T values from the first-order t-tests comparing rate in
the [0;500ms] post-cues onset epoch to baseline rate ([-500;0ms] epoch before cues onset).
*: p<0.05; ***: p<0.001
(b) LC neurons positively encoded immediate
subjective difficulty to engage in choice around cues
onset
Raster examination indicated that some LC neurons displayed higher activity around
cues onset when RT (time from cues onset to action onset) would be longer (Figure 60A). For
a large proportion of the neurons with such a pattern of activity, the effect appeared to start
before cues onset (about 100ms before). I first quantified the effect of total RT on LC activity
in a [-100;200ms] epoch around cues onset. I fit a GLM for spike count in that epoch with only
RT as parameter. This analysis was restricted to non-repeated trials in which the monkey later
engaged in the choice. Additionally, in order to be able to decompose RT into its
subcomponents, I focused on the 50 neurons for which eye-tracking was available. The effect
of RT on peri-cues onset spike count was significant for 9/50 neurons (1 negative, 8 positive),
and was significantly positive at the level of the whole population (t(49)=2.07; p=0.044, Figure
60B).
In order to specify the relationship between LC activity and the subcomponents of RT,
i.e. action-selection and action-triggering time, I then fit a GLM for spike count (same window,
same neurons, same trials) with action-selection and action-triggering times as parameters.
The encoding of action-selection time was significant for 10/50 neurons (9 positive, 1
negative), and was significantly positive at the level of the population (t(49)=2.02; p=0.049,
Figure 60C). The encoding of action-triggering time was significant for 7/50 neurons (6
positive, 1 negative), and was significantly positive at the level of the population (t(49)=2.09;
p=0.042).
To verify whether the effect of the two parts of RT were encoded over and above the
effects of offer parameters (Forces and Rewards), I then added chosen and unchosen Forces
and Rewards to the previous GLM. In that model, the effect of action-selection time was still
significant for 10/50 neurons (1 negative, 9 positive) but there was only a marginally significant
Experimental results
220
tendency for a positive encoding at the level of the whole population (t(49)=1.70, p=0.095,
Figure 60C). The encoding of action-triggering time was still significant for 7/50 neurons (6
positive, 1 negative) and was consistently positive across the population (t(49)=2.74,
p=0.0085). This pattern indicates that neurons encoded the subjective difficulty to engage in
the choice and trigger the chosen action.
Figure 60: LC neurons encoded RT positively, and specifically its subcomponent action-
triggering time, over and above offer parameters, around cues onset.
(A) Raster plot of the activity of an example LC unit around cues onset (green line). Magenta
dots represent the activity of the neuron when the monkey will choose the left side option, and
purple dots represent the activity of the neuron when the monkey will choose the right side
option. Diamond-shaped purple dots represent the timing of the onset of the side press. Around
cues onset, this neuron exhibited increased activity when the monkey would take longer to
respond (longer time between green line and purple diamond).
Experimental results
221
(B,C,D) GLMs for spike count in the [-100;200ms] peri-cues onset epoch, with (B) only RT, (C)
action-selection and action-triggering time, or (D) action-selection, action-triggering time and
chosen and unchosen offer parameters (Forces and Rewards).
(B) LC neurons encoded RT positively around cues onset.
(C) LC neurons encoded action-selection and action-triggering times around cues onset, but (D)
only action-triggering time was significantly encoded over and above effects of Forces and
Rewards.
*: p<0.05, **: p<0.01; n.s.: non significant, F: Force, R: Reward, ctt: constant
In order to confirm that, I tested whether LC neurons encoded the current choice
engagement rate around cues onset. First, I split sessions in halves, determining the choice
engagement rate in each half, and associating each trial with the current choice engagement
rate. Then, selecting only non-repeated trials in which the monkey had engaged in the choice,
I fit neuron-by neuron GLMs for spike count with only the current choice engagement rate as
a parameter. The effect of the current engagement rate was significant for 14/50neurons (6
negative, 8 positive) and the effect was not consistent across the population (t(49)=0.079;
p=0.94)19. Neurons did not encode middle press RT or current engagement rate in the middle
press around that time either20. Importantly, neither side press RT, nor action-triggering time
were encoded in the baseline activity (activity prior to middle red-dot onset or earlier before
cues onset, i.e. before 100ms before cues, during middle blue dot presentation).
Similarly to what I observed around middle press onset, LC neurons therefore encoded
trial-by-trial difficulty to engage in choices (action-triggering time for side press, middle press
RT for middle press). In other words, LC neurons encoded the subjective difficulty of ongoing
action-triggering processes. Conversely, they did not encode the current engagement rates in
middle presses and choices. Considering that engagement rates provide long-term
measurement of the difficulty to engage, i.e. assess motivational states, the fact that they
were not encoded by LC neurons would be in favor of an encoding of the subjective difficulty
of ongoing processes with a fine time scale.
19 Adding action-selection and action-triggering time and chosen Forces and Rewards: current choice engagement significant for 12/50 neurons (6 positive, 6 negative), population: t(49)-0.10; p=0.92. 20 Middle press engagement rate: significant for 14/50 neurons (5 negative, 9 positive), population: t(49)=1.36; p=0.17. Middle press RT: significant for 5/50 neurons (1 negative, 4 positive), population: t(49)=0.69; p=0.49.
Experimental results
222
Around cues onset, contrary to what I found around middle press onset when LC
neurons encoded RT just prior to action onset, and to what we observed in the first study
(Bornert and Bouret, in prep) LC neurons encoded action-triggering time early before action
onset. Indeed, most RTs were longer than 700ms, meaning that the window in which action-
triggering time was encoded anticipated the action by at least 500ms21.
(c) LC neurons encoded Rewards but not Forces
after cues onset
Next, I questioned whether LC neurons encoded any of the parameters (Forces and
Rewards) of the offers displayed on screen, from which the monkey could choose. For activity
to participate in offer value evaluation, LC neurons would have needed to encode both
dimensions of both offers, i.e. the Forces requested and the offered Rewards. I restricted our
analysis to trials in which the monkey later made a choice i.e. pressed on one of the two side
grips. In the previous part, I already showed that LC neurons encoded only the chosen Reward
early after cues onset (Figure 60D).
Based on raster examination (rasters for one example unit displayed in Figure 61A&B),
I assessed effects of chosen and unchosen Rewards and Forces on LC activity in a [200;500ms]
epoch after cues onset. Although this window was late after cues onset, I ensured that I did
not capture pre-action effects by only analyzing trials with a total RT above 700ms (although
identical results were obtained with trials with shorter RTs included), and by verifying that
adding RT to the model did not capture the effects of offer parameters. I fit a neuron-by-
neuron GLM for spike count in that epoch with chosen and unchosen Forces and Rewards as
parameters. Then, I tested for population effects of each parameter by performing a second-
order t-test on the regression coefficients obtained.
The effect of the chosen Reward was significant for 7/54 neurons (6 positive and 1
negative), and was consistently positive at the level of the population (t(53)=3.21; p=0.0022,
Figure 61C). The effect of the chosen Force was significant for 2/54 neurons (1 positive and 1
negative) and was not significant at the level of the population (t(53)=-1.60; p=0.11). As for
21 Note that I could not find any later encoding of side press RT as a whole or of action-triggering time, even just before action onset. [-300;0ms] from side press onset, GLM for spike count with only action-triggering time as parameter: effect of action-triggering time significant for 10/50 neurons (5 positive and 5 negative), not significant across the population: t(49)=0.63; p=0.53.
Experimental results
223
the unchosen parameters, the effect of the unchosen Reward was significant for 5/54 neurons
(2 positive, 3 negative) and was marginally significant at the level of the population
(t(53)=-1.94; p=0.058). The effect of the unchosen Force was significant for 4/54 neurons (all
positive) and was not significant at the level of the population (t(53)=-0.50; p=0.62). These
results indicate that LC neurons did encode some of the information taken into account for
choices between alternatives, but not all of it. Indeed, in our population of LC neurons, I did
not find any consistent encoding of the Force parameters after cues onset, while there was an
encoding of the Rewards.
(A) Trials sorted by chosen Reward (B) Trials sorted by unchosen Reward
(C)
Figure 61: Encoding of offer parameters (Forces and Rewards) in a late epoch ([200:500ms])
after cue onset: LC neurons only encoded Rewards.
(A,B) Rasters of two example LC units showing the contrast of activity between different (A)
chosen or (B) unchosen Reward conditions. Activity is aligned on cues onset (vertical green
line). Magenta dots represent spikes when the monkey later chose the left side offer and purple
Experimental results
224
dots represent spikes when the monkey later chose the right side offer. Purple diamond-shaped
dots represent action onset. Note also that the LC neuron in (A) encoded RT positively after
cues onset, i.e. displayed more spikes just after cues onset when RT (green line to purple
diamond interval) would be longer.
(C) Summary of neuron-by neuron GLMs for spike count in the [200;500ms] epoch after cues
onset. Bars represent the mean coefficients for each parameter, across neurons, with chosen
and unchosen Force and Reward parameters. At the level of the whole population, there was
a significant positive encoding of the chosen Reward and a marginally significantly negative
encoding of the unchosen Reward. There was no significant encoding of the chosen and
unchosen Forces.
**: p<0.01; n.s.: non significant, F: Force, R: Reward, ctt : constant
(d) Increased activity of LC neurons around cues
onset was associated to increased probability to choose
high-Force option
Next, I questioned whether the activity of LC neurons around cues onset influenced
behavior, and especially choices. In Borderies et al., (2020), we showed that decreasing brain
NA led to a stronger effect of Effort on choices and overall less high-Effort choices. Here, I
tested whether I could find any signature of such an effect in LC activity around cues onset,
although I could not verify any causality link. In other words, I tested whether LC activity
around cues onset could predict the sensitivity of choices to Forces and/or the overall
propensity to choose high-Force offers. I conversely expected no relation between LC activity
and Reward-based decisions.
Using the same method as before (see II.B.3.a)(2)(e)), I separated choices between
high-Force versus low-Force or high-Reward versus low-Reward choices and fit logistic
regressions for high-Force or Reward choices. First, I fit a logistic regression for high-Force
choices, with the difference in Force (high minus low Force) and in Rewards (Reward of high-
Force option minus Reward of low-Force option), the interaction between the difference in
Forces and action-selection times, and spike count in the [-100;200ms] epoch around cues
onset. Over and above previously described effects22, there was a positive effect of spike count
22 Difference of Forces: β=-0.62, p<10-14; Difference of Rewards: β=4.26, p<10-124; Interaction between action-selection time and difference in Forces: β=-0.16, p=0.039
Experimental results
225
(β=0.14; p=0.048, Figure 62A). The interactions between spike count, action-selection time
and differences in Forces or Rewards were not significant if added to the model (p>>0.05).
Figure 62: LC activity around cues onset had a positive influence on the probability to choose
high-Force options, on the exerted force, especially in high-chosen Force conditions, and a
positive effect on performance.
Experimental results
226
(A) Logistic regression for high-Force choices (1 if the high-Force option was chosen, 0
otherwise). Over and above effects of the difference of Forces (high – low), of the difference of
Rewards (Reward of the high-Force option – Reward of the low-Force option), and of their
interaction with action-selection time, spike count in the [-100;200ms] peri-cues onset epoch
had a positive effect on the probability of high-Force choices.
(B) GLM for the maximum exerted force on the side grip. Over and above effects of offer Forces
and Rewards, chosen side of press, and action-triggering time, there was a positive effect of
spike count on the exerted force and a positive effect of the interaction between spike count
and the chosen Force.
(C) Logistic regression for performance of the side press. Over and above effects of offer Forces
and Rewards, side of press, and action-triggering time, there was a marginally significant
positive effect of the interaction between spike count and the chosen Force.
(C) Interaction plot showing the effect of the interaction between chosen Force and spike count
at cue onset (x-axis) on performance (y-axis). Spike count is split between below and above the
median (m).
*: p<0.05, ***: p<0.001; n.s.: non significant, F: Force, R: Reward, ctt : constant; Spk: spike.
Using the same procedure, I fit a logistic regression for high-Reward choices, with the
difference in Rewards (high-low Rewards) and in Forces (Force of high-Reward option – Force
of low-Reward option), and spike count in the [-100;200ms] epoch around cues onset. Over
and above previously described effects23, the effect of spike count was marginally significant
(β=0.21; p=0.073). There was no effect of the interactions between spike count and other
parameters (p>>0.05).
Importantly, the main effect of spike count on high-effort choices probability was not
present in the baseline ([-400;-100ms] pre-cue or [-300;0ms] pre-red dot onset), in the pre-or
peri-middle press epoch ([-200;0ms] and [0;200ms]) or in the pre- or post-action onset epoch
of the previous press ([-200;0ms] and [0;600ms] of the last trial with a press). This indicates
that this relation between LC activity and high-Force choices was specifically associated to
cues anticipation and processing. LC activity around the onset of the cues was associated to a
23 Difference of Forces: β=-1.90, p<10-39; difference in Rewards: β=1.51, p<10-24
Experimental results
227
higher probability to choose the high-Force options, but no difference in effort sensitivity. This
suggests that higher LC activity may have facilitated cognitive control exertion for the choice,
decreasing the bias towards avoiding high-Force offers. Conversely, there was no significant
relation between cues onset activity and Force-sensitivity (no interaction between Force
difference and spike count).
(e) Activity around cues onset had a positive effect
on side press exerted force in high-chosen Force
conditions
Next, I questioned whether this relationship between high-effort choices and peri-cues
activity could be accounted for by an increased capacity or willingness to perform the forces
on the side grip and therefore an increase in performance following increased LC activity
around cues onset. More specifically, I expected either one of two results: either increased
cues onset activity would lead to overall stronger presses, regardless of the force category,
i.e. spike count would have a global energization effect on side presses (main effect on exerted
force), or it would optimize the exerted force for the requested level of force (interaction with
chosen Force).
I pooled all the trials from all the recording sessions (with valid eye-tracking, 50
neurons) in which the monkey had pressed on the middle grip then made a choice (pressed
one of the side grips) and in which I could identify the time of the choice between offers (end
of action-selection time). I fit a GLM for the exerted force on the side grip. I used peri-cues
spike count ([-100;200ms]), action-triggering time, chosen and unchosen Forces, Side of press,
and the interaction between spike count and chosen Force as parameters. Besides identical
effects of offer parameters, side of press and action-triggering time as described earlier24,
there was a main effect of spike count (β=0.021; p=0.028) and positive interaction between
spike count and chosen Force25 (β=0.022; p=0.016, Figure 62B). This indicates that pre-middle
press spike count both had non-specific effects on the exerted force during the side press, and
chosen-Force specific effects, increasing the exerted force most strongly in conditions where
stronger force exertion was required.
24 Chosen Reward: β=0.045; p<10-4; Chosen Force: β=0.67; p<10-150; Unchosen Reward: β=-0.054; p<10-6; Unchosen Force: β=0.050; p<10-6; Action-triggering time: β=-0.040; p<10-4; Side of press: : β=0.44; p<10-112. 25 No other interaction was significant when included.
Experimental results
228
Consequently, I looked for effects of peri-cues onset spike count on performance.
Given that performance in case of chosen Force 1 was virtually at ceiling, I did not expect any
main effect of spike count on performance, but only when the monkey chose the Forces 2 or
3. Thus, I fit a logistic regression for performance with chosen and unchosen Forces and
Rewards, action-triggering time, side of press and the interaction between spike count and
the chosen Force as parameters. Over and above effects of offer parameters and action-
triggering time26, there was a marginally significant positive effect of the interaction between
spike count and the chosen Force on performance (β=0.039; p=0.069, Figure 62C&D).
The effect of the interaction between spike count and chosen Force on the exerted
force were already present before middle red dot onset ([-300;0ms]), in the baseline rate
(β=0.0099; p=0.013), but not the main effect of spike count (β=0.013; p=0.15). Since this
similarity could have been due to the strong correlation between baseline rate and rate
around cues onset (R=0.48; p<10-280), I tested whether a model using the baseline spike count
rather than the spike count around cues onset fit the data better. The model including the
baseline rate had a higher Bayesian Information Criterion (BIC=9915.3) than the model
including the rate around cues (BIC=9907.5). Moreover, adding the interaction between the
baseline spike count and the chosen Force to the model including the cues onset spike count
and its interaction with the chosen Force did not improve the fit of the model, as its BIC
increased (BIC=9919.4). Consequently, the relationship between cues onset spike count and
the exerted force on the side grip were not merely due to effects of baseline rate on the
exerted Force.
Regarding activity before and during side press ([-300;0ms] and [0;300ms]), the
interaction between spike count and the chosen Force was significantly positive, yet very small
(before press: β=0.0075; p=0.036; during press: β=0.0078; p=0.023) and there was no main
effect of spike count (before: β=-0.012; p=0.21; during: β=-0.0021; p=0.83). However, the fits
of these models were not as good as the fit of the one including cues onset rate instead (BICpre-
action rate=9918.3; BICperi-action rate=9918.4; BICperi-cues rate=9907.5), and adding the interaction
between pre-action rate and the chosen Force to the model with cues onset rate did not
improve model fit (BICpre-action rate=9921.2; BICperi-action rate=9919.7).
26 Chosen Reward: β=0.16; p=0.019; Chosen Force: β=-1.87; p<10-89; Unchosen Reward: β=-0.16; p=0.018; Unchosen Force: β=0.23; p=0.00015; Action-triggering time: β=-0.23; p<10-4; Side of press: : β=0.05; p=0.62.
Experimental results
229
Note also that such effects of the interaction between spike count and chosen Force
on the exerted force were not present for the spike count around the middle press (before,
[-200;0ms] or after, [0;200ms]), earlier during the inter-trial interval ([-450;-150ms] from
middle red dot onset), or later after cues onset ([200;500ms] after cues onset).
Thus, LC activity around cues onset had a chosen-Force dependent positive effect on
the exerted Force. In other words, increased LC activity around the time of offer presentation
predicted increased force exertion for the chosen action, especially in difficult conditions,
resulting in a tendency for an increase in performance in difficult conditions only. This suggests
that increased LC activity around cues onset facilitated physical effort production for the
chosen action, better adjusting the force exerted to the condition of the task. Indeed, in the
small Force condition (Force 1), in which almost every attempt to press the grip was successful,
increasing the exerted force was not instrumental, as it would not be beneficial for
performance. Conversely, in the intermediate Force condition (Force 2), and even more in the
high Force condition (Force 3), increasing the force increased the probability of success (and
of reward delivery).
These findings provides a potential mechanistic account for the relation between LC
activity and force sensitivity (Borderies et al., 2020), as it suggests that producing an effort (at
least physical effort) may have been less difficult after higher LC activity. This idea resonates
with the results presented above (see parts II.B.3.b)(1)(c) and II.B.3.b)(2)(d)) regarding the
relation between LC activity around the middle press and the subjective difficulty to engage
in the choice and trigger the chosen action, as well as the finding that increased LC activity
around cues onset led to a decreased bias against choosing high-Force offers. Indeed, these
findings suggest that following higher LC activity, the monkey found it easier to produce
cognitive control to repress his natural tendency to disengage from the task, and to inhibit the
bias towards avoiding high-effort offers.
Experimental results
230
(3) Side-press related LC activity
(a) Evoked activity around action onset
Various studies have reported that LC neurons activate just prior to action onset
(Bouret et al., 2012; Bouret and Richmond, 2015, 2009; Bouret and Sara, 2004; Clayton et al.,
2004; Varazzani et al., 2015, Study 1, see part II.A), and maintain higher activity during
sustained actions such as presses on force grips (Varazzani et al., 2015). Here, inspection of LC
activity during the second step of the task revealed that LC neurons displayed bursts of activity
(a few spikes) between choice onset (gaze locking onto chosen option) and side press onset,
and/or during side press. I examined whether I could identify if activity between choice onset
and side action onset was better aligned onto choice onset or side action onset, using the
same method as the one used for activity between cues onset and choice onset. There was a
tendency for a better alignment of spikes onto the onset of the press rather that of the choice
(t(49)=1.77; p=0.83). Identically, during the side press, we examined whether we could
identify if LC spikes were better aligned onto the onset of the press or the attainment of the
maximum of exerted force. For all 54 neurons, after identifying the time of the peak of force
for each non-repeated trial in which the monkey made a side press, we compared the timings
between press onset and spikes, or spikes and time of peak of force. Across the population,
the comparison was not significant (t(49)=1.30; p=0.20).
We then quantified the proportion of neurons that responded to side action onset,
using the same technique as the one used for middle press and cues onset (neuron-by-neuron
t-tests to compare rate in window of interest to baseline rate, then second order t-test on T
values obtained for population effects). Before side action onset, in a [-300;0ms] epoch, we
compared rate to activity in a [-300;0ms] epoch before cues onset. 35/54 neurons exhibited a
change in rate (26 increased, and 9 decreased, Figure 63A), and there was a coherent increase
in rate across the population (second-order t-test: t(53)=2.56; p=0.013, Figure 63B). During
action execution, in a [0;300ms] from side action onset, 47/54 neurons exhibited a change in
rate (36 increased and 11 decreased, Figure 63A), and at the level of the whole population
there was a consistent increase in rate (t(53)=3.53; p<10-3, Figure 63B).
Experimental results
231
(A) (B)
Figure 63: Rate of LC neurons around side action onset, and results of first and second order t-
tests assessing rate changes.
(A) Activity of the 54 recorded LC units around action onset. Rate is z-scored by neuron. The
green dots indicate the mean timing of the onset of the cues by session i.e. by neuron.
(B) Density plots (bandwidth=5) of the T values from the first-order t-tests comparing rate in
the [-300;0ms] pre-action onset epoch and in the [0;300ms] post-action onset epoch to
baseline rate ([-300;0ms] epoch before cues onset).
(b) LC neurons encoded the difficulty of the action
during its execution
Based on the literature (Varazzani et al., 2015) and on what we found in our first study,
we expected LC neurons to encode the immediate difficulty to perform the action while the
monkey performed it. Practically, we expected LC neurons to encode the chosen Force and
the maximum exerted force during the press, and such an effect was visible on raster plots
(Figure 64A). We first fit a GLM with only offer parameters (chosen and unchosen Force and
Reward) for spike count in the [0;600ms] epoch after action onset, i.e. during the press (see
Figure 48 for time course of force exertion). Once again, we only used non-repeated trials in
which the monkey made a side press. Since eye-tracking was not required, all 54 neurons were
included. The effect of the chosen Force was significant for 22/54 neurons (20 positive, 2
Experimental results
232
negative) and was consistently positive across the population (t(53)=5.41; p<10-5, Figure 64B).
The effect of the chosen Reward was significant for 4/54 neurons (2 positive, 2 negative) and
was not consistent across the population (t(53)=0.41; p=0.69, Figure 64B). The effect of the
unchosen Reward was significant for 5/54 neurons (3 positive, 2 negative) and was
consistently negative across the population (t(53)=-2.35; p=0.023, Figure 64B). The effect of
the unchosen Force was significant for 4/54 neurons (2 positive, 2 negative) and was not
consistent across the population (t(53)=1.27; p=0.21, Figure 64B). Note that LC neurons
already encoded the chosen Force before action onset27.
We then fit a GLM for spike count in the same epoch with only the maximum exerted
force on the side grip as parameter. The effect of the maximum exerted force was significant
for 19/54 neurons (2 negative, 17 positive) and was consistently positive across the population
(t(53)=4.27; p<10-4, Figure 64C). When we added chosen and unchosen Forces and Rewards
to the model, the effect of the maximum exerted force was still significant for 13/54 neurons
(7 negative, 6 positive) and no longer significant across the population (t(53)=0.97; p=0.33,
Figure 64D).
Finally, in order to provide some conclusion on whether LC neurons encoded the
objective difficulty (chosen Force category) of the action or the actual resource engagement
in the action, we compared the fit of the model including only the chosen and unchosen Forces
and Rewards to the fit of the model including only the exerted force. For 24/54 neurons, the
model with the maximum exerted force alone fit the date better, and for 30/54 neurons, the
model with the chosen Force fit the data better. In these conditions, we cannot properly
conclude on whether LC neurons as a population encode one rather than the other. Our
results only provide an partial indication that LC neurons encode the difficulty of chosen
actions during their execution, at least the objective difficulty (required force).
27 [-250;0ms] epoch before action onset: Chosen Force: significant for 11/54 neurons (3 negative, 8 positive), consistently positive across the population (t(53)=2.56, p=0.013). Chosen Reward: significant for 6/54 neurons (4 negative, 2 positive), not consistent across the population (t(53)=-0.50; p=0.62). Unchosen Reward: significant for 4/54 neurons (1 negative, 3 positive), not consistent across the population (t(53)=1.04; p=0.30). Unchosen Force: significant for 3 neurons (1 negative, 2 positive), not consistent across the population (t(53)=1.24, p=0.22). NB: GLM with only action-triggering time, significant for 10/50 neurons (5 positive and 5 negative), not significant across the population (t(49)=0.63; p=0.53).
Experimental results
233
Figure 64: LC neurons encoded the difficulty of the action during its execution.
(A) Raster plot of the activity of an example LC unit around the time of action execution (around
action onset). Trials are split by chosen Force (F1, 2 or 3, top to bottom). Purple dots indicate
spikes in trials in which the left side offer was chosen and pink dots indicate spikes in trials in
which the right side offer was chosen. The pink line indicates action onset. The green dots
indicate cues onset. This unit was more active during the press when a higher Force was chosen
and executed.
(B-D) Summary (mean and SEM) of the coefficients of the neuron-by-neuron GLMs for spike
count with (B) chosen and unchosen Forces and Rewards, (C) the maximum exerted force, or
(D) all the latter as parameters. (B) LC neurons encoded the chosen Force and the unchosen
Reward negatively. (C) When the maximum exerted force was alone in the model, it had a
positive effect on spike count, but (D) this effect was not significant over and above the effects
of task parameters.
F1
F2
F3
Experimental results
234
(c) Higher press-related LC activity at trial N was
associated to increased Force sensitivity of choices at
trial N+1
Next, we questioned whether the activity of LC neurons around the time of the press,
at the time when they encode the exerted force, and therefore putatively the physical effort,
could have an influence on the choice in the next trial. Specifically, if LC neurons participate in
the monitoring of effort exertion to subsequently adjust effort allocation, we would expect
that increased LC activity at the time of the exertion, putatively reflecting the resources
invested in the action, would lead to increased effort sensitivity in the next trial.
To test that hypothesis, in order to specifically focus on effort-based choices, we
focused on trials in which the monkey made a choice (made a side press) and had made a
choice in the previous trial, and in which the two offers presented had identical Rewards, i.e.
did not differ in the Reward dimension, but differed in the Force dimension. We split trials
between high-Force choices (the offer with the highest Force proposed was chosen) or low-
Force choices (the offer with the lowest Force proposed was chosen). We fit a logistic
regression for high-Force choice (1 if the monkey made a high-Force choice, 0 if he made a
low-Force choice) with the difference in Forces offered (high minus low), action-selection
time, the Force exerted in the previous trial, and the interaction between spike count in the
previous trial around side press ([0;600ms] from press onset) and the difference in offered
Forces as parameters.
There was a significant negative effect of the interaction between spike count during
the previous press and the Force difference (β=-0.30; p=0.037, Figure 65A), indicating that
increased spike count around the previous press increased the Force sensitivity of Force-based
choices in the next trial (Figure 65B). Besides this effect, the difference in Forces had a negative
effect (β=-0.62; p=0.00020) indicating that the monkey was more reluctant to choose the high-
Force option if it had a much higher Force requirement than the alternative option. Moreover,
there was a positive effect of action-selection time (β=0.28; p=0.013, Figure 65A) indicating
that longer action-selection times were associated with higher probability to choose the high-
Force offer. Finally, the maximum exerted force of the previous trial had a positive effect
(β=0.81; p<10-11, Figure 65A). Importantly, the effect of the interaction between spike count
and Force difference was also true when considering only trials following correct execution of
Experimental results
235
the side press, meaning that it was not due to the increased probability of failure for high-
Force conditions.
The effect of the interaction between spike count and the Force difference was also
present when considering spike count around cues onset ([-100;200ms] from cues onset in
the same trial as the choice, β=-0.30; p=0.010), but the model using the spike count related to
the previous press fit the data considerably better (BICcues = 735.49; BICprevious press=484.8). I
could not find any evidence for such an effect at any other time during the trial or in the
baseline (inter-trial interval) rate.
Figure 65: Increased spike count during the previous side press was associated to increased
Force sensitivity of the next Force-based choice.
(A) Coefficients of the logistic regression for Force-based choices (offers with identical Reward
size and different Forces). Over and above effects of the Force difference and action-selection
time of the current trial, and of the exerted Force of the previous press, the interaction between
the spike count of the LC neurons during the execution of the previous chosen action and the
difference of Forces (high minus low Force) had a negative effect on the probability to make
high-Force choices.
(B) Interaction plot showing the interaction between the effect of the Force difference (x-axis)
and the spike count during the previous press (split between spike count above, dotted line, or
Experimental results
236
below median spike count m, full line) on the probability to make high-Force choices (y-axis).
When spike count during the previous side press increased, the effect of the difference in Force
was stronger, indicating that Force sensitivity increased. *: p<0.05; **: p<0.01; ***: p<0.001;
n.s.: non significant; m: median;
Altogether, these results indicate that the spike count of LC neurons during the chosen
action at trial N scaled with the difficulty and possibly the effort produced for action execution.
Moreover, increased spike count around that time was associated to increased Force
sensitivity of choices at trial N+1. In other words, LC activity during the press possibly reflected
the physical effort produced to execute it, and increased activity in that period of time was
associated to increased physical effort avoidance the next time the monkey faced a binary
choice. One of the working hypotheses of this study was that the activity of LC neurons during
effort production could serve as a feedback for the effort allocating system, used to adjust
future effort-based decisions. In order to regulate resource expenditure, following intense
effort production, the system should favor options requiring less effort. Our findings show
that LC activity during actions (at least actions requiring physical effort production) could
provide information about resource expenditure, and is associated to modulations of effort
allocation strategies. Thus, this data suggests that effort-related LC activity could be
responsible for adjusting effort sensitivity for subsequent choices based on the a posteriori
evaluation of effort exertion.
(4) Inter-trial activity of LC neurons
(a) LC activity increased during the ITI following
failures to execute the side press
LC activity has been proposed to participate in action-monitoring and especially error
monitoring. Jahn and colleagues (2020) reported increased LC activity following errors
(fixation breaks and therefore no engagement), scaling with the probability that the monkey
would re-engage in the task, suggesting that LC activity signals errors and subsequently
enhances resource mobilization for engagement. I aimed at extending this finding beyond
participation mistakes (failures to engage). If LC neurons do track mistakes, in this task, I would
expect LC neurons to increase their activity after failures to engage and after failed presses.
Experimental results
237
ITIs were 475±25ms long, and, in the case of successful trials, followed reward delivery. Based
on raster inspection (Figure 66A), I used a [-450;-150ms] epoch from the onset of the next trial
(middle red dot) to analyze ITI activity, in order to avoid any confusion with Reward delivery-
related activity.
(A) Activity during ITI
Figure 66: LC neurons encoded performance of previous side press, but not engagement in
previous side press during the ITI ([-450;-150ms] before middle red dot onset).
Experimental results
238
(A) Raster plot of the activity of an example LC unit during the ITI, showing only ITIs following
engagement in choices, and separated based on the outcome of the side press (failure, top, or
success, bottom).
(B,C,D) Summary of the neuron-by-neuron GLMs for spike count during the ITI.
(A) LC neurons as a population did not significantly encode engagement in the choice in the
previous trial (N-1) during the ITI.
(C) LC neurons encoded performance in the previous trial (N-1) negatively, i.e. were more active
after a failure.
(D) LC neurons encoded performance in the previous trial (N-1) negatively, and there was a
marginally significant effect of the interaction between performance and chosen Reward of
the previous trial.
*: p<0.05; n.s.: non significant, ctt : constant
Regarding ITIs following failure to engage, I focused on ITIs following trials in which the
monkey had made a middle press, and separated them between trials in which the monkey
made a side press (choice) or not. Trials in which no middle press was initiated were excluded.
I fit a GLM for spike count in the ITI window ([-450;-150ms] from middle red dot onset) with
engagement in the previous side press as parameter (0 if no press was made on the side, 1
otherwise, regardless of performance of side press). I had to exclude 7/54 neurons due to
insufficient number of trials in which the monkey did not engage in the choice after
performing the middle press. The effect of engagement in the previous choice was significant
for 10/47 neurons (9 negative, 1 positive). There was no effect of engagement in choice in the
previous trial on ITI activity at the level of the population (second order t-test: t(46)=-0.90;
p=0.37, Figure 66B).
Regarding ITIs following failure to perform the side press, I focused on ITIs following
attempts to perform the side press, and split them between ITIs following failure or success in
performing the side press. I then fit neuron-by-neuron GLMs on ITI spike count with
performance of the side press (1 if success, 0 if failure) as parameter. The effect of
performance of the previous trial was significant for 15/54 neurons (13 negative, 2 positive),
and the effect was consistently negative across the population (second order t-test: t(53)=-
1.87; p=0.047), indicating that LC neurons displayed increased spike count after a failed side
press (Figure 66A&C).
Experimental results
239
This encoding of performance could, in fact, correspond to an encoding of Reward
prediction error. Indeed, there would be a high prediction error in case of failed presses,
especially if the missed Reward was big. To test that, I added the interaction between
performance and Reward at stake for the previous press to the previous model. The effects of
performance was still significant for 15/54 neurons (13 negative, 2 positive) and significant
across the population (second order t-test: t(53)=-2.03; p=0.048, Figure 66D). The effect of the
interaction between the Reward and performance of the previous trial was significant for
17/54 neurons (12 negative, 5 positive) and was marginally significantly negative at the level
of the population (t(53)=-1.85; p=0.071). This indicates that the relation between LC activity
during the ITI and performance was not confounded with an encoding of Reward prediction
error. Considering the marginally significant result obtained for the interaction between
performance and Reward at stake, it cannot be excluded that LC neurons do encode Reward
prediction errors, although this would deserve further investigation.
Overall, despite not replicating the finding by Jahn and colleagues (2020) regarding the
response of LC neurons to commitment errors, I showed that LC neurons discriminated
between correct and incorrect performance, with a higher spike count after incorrect
performance. In other words, in this task, LC neurons responded to execution errors.
Discussion
240
III. Discussion
The work conducted during my PhD aimed at clarifying the role of the LC/NA system in
effort processing. We hypothesized that LC neurons provide an online and domain-general
representation of effort production, and signal some of the consequences of effort allocation,
or at least insufficient effort allocation, such as errors. We proposed that these modulations
of activity could serve as a feedback signal for structures making the decisions to allocate
effort (cognitive control and physical effort). Moreover, we hypothesized that LC activity had
beneficial short-term effects on effort production.
In three tasks involving investing effort, physical effort and/or cognitive control, we
consistently showed that LC neurons encoded the difficulty of engaging in the actions and the
difficulty of executing them. More precisely, before action onset, LC neurons encoded RT, i.e.
the time taken to respond by triggering the actions, which, at least in the Delay-discounting
and two-alternative Force-discounting task, scaled negatively with the subjective cost of
engaging in actions. Additionally, during the action itself, LC neurons encoded the force
exerted by the monkeys in the two tasks involving differential force exertion.
In the second study, I showed that LC activity around costly task events (engaging in
actions and executing them) had delayed effects that could be separated into two
subcategories: short-term within trial effects, and long-term across trials effects. First, in
periods of time in which LC neurons encoded the subjective difficulty to engage in the current
action, LC activity had delayed within-trial facilitating effects on future effort production,
which could be accounted for by a decrease in the subjective cost of engaging in future actions
and executing them. This effect was domain-general, meaning that increased LC activity was
associated to enhancements of cognitive control and physical effort production. Second,
across-trials, increased LC activity around the time of exertion of the physical effort (grip press)
was associated with increased sensitivity of the following choice to the effort dimension,
suggesting that it served as a learning signal for structures participating in effort-based
decision-making.
In this part, I discuss the findings of the two studies detailed in previous parts of this
manuscript, their implications with regards to the literature of effort allocation and production
and of LC neurophysiology, as well as potential perspectives for future studies.
Discussion
241
A. Behavior: signatures of effort in one- and two-
alternative tasks
In the two studies detailed earlier, three different tasks manipulating physical and
mental effort were used: one-alternative Force-discounting and Delay-discounting tasks, and
a two-alternative Force-discounting task. In the one- and two-alternative Force-discounting
tasks, high-Force options were perceived as aversive, i.e. were avoided more often than low-
Force options. Indeed, higher Forces led to decreased acceptance rates (e.g. Jahn et al., 2020;
Varazzani et al., 2015; Zénon et al., 2014) and were chosen less often (e.g. Borderies et al.,
2020; Cousins and Salamone, 1994; Floresco et al., 2008; Salamone et al., 2007; Stevens et al.,
2005; Walton et al., 2006). This pattern followed the law of least effort (Hull, 1943). In the
Delay-discounting task, high-Delay options were refused more often (Falcone et al., 2019;
Minamimoto et al., 2012, 2009b), as would be expected by a discounting of Reward values by
Delay (e.g. Green et al., 2013; Kirby and Maraković, 1996; Madden et al., 2003; McKerchar et
al., 2009; Prévost et al., 2010; Rachlin and Jones, 2008; Shamosh and Gray, 2008). Overall, our
results indicate that monkeys perceived high-cost options as more aversive.
1. Cognitive control exertion for decisions
Goal-directed behavior relies on the representation of the relation between actions
and their outcomes, i.e. of subjective utility (Balleine and Dickinson, 1998; Dickinson and
Balleine, 1994; Valentin et al., 2007; Zwosta et al., 2015). Cognitive control serves to maximize
subjective utility (Kool and Botvinick, 2018). In the particular case of decision-making, exerting
control allows overcoming biases arising from the use of automatic processing (Shiffrin and
Schneider, 1977) and therefore overriding default responses (Epstein, 1994; Sloman, 1996).
This role was particularly emphasized by pupillometry studies showing that increased pupil
diameter, a proxy for the invested cognitive control (Alnæs et al., 2014; Wahn et al., 2016;
Wright et al., 2013), was associated to decreased decision biases (de Gee et al., 2020, 2017,
2014).
In our three effort-discounting tasks, a first step at which behavior had to be adjusted
in order to maximize utility was the engagement in trials. Monkeys could refuse to engage in
trials based on cues information, although rejected options would be offered again until they
Discussion
242
were completed. Refusing trials led to lower reward rates, and therefore lower utility28, as it
led to increased delay to reward delivery. This was especially true in the two-alternative Force-
discounting task, which had longer trials since they included a middle press before offers were
made. Considering that high-cost/low-reward offers were perceived as aversive, the
automatic behavior would have been disengaging from the task when facing low-value offers.
Since cognitive control is required to override decision biases and automatic behavior,
choosing to engage in such trials probably required exerting some cognitive control. Engaged
trials could therefore be seen as trials during which sufficient control was allocated such that
the monkey decided to commit. Nonetheless, in the absence of formal evidence that decisions
to engage were goal-directed, it cannot be considered as certain. Indeed, it could still have
been an automatic Pavlovian approach behavior (Dickinson, 1980; Timberlake, 1994).
In the second study, besides engaging in trials, maximizing utility involved making
choices that minimized force exertion while maximizing rewards. Behavioral analyses showed
that the monkey’s choices were compatible with a cost-benefit trade-off. Still, rather than
engaging cognitive control and performing extensive computations to decide optimally, he
could have relied on simplifying decision strategies (Gigerenzer, 2010; Gigerenzer and
Goldstein, 1996; Sidarus et al., 2019; Tversky and Kahneman, 1974), such as perceptual
strategies (locating the offer with the biggest reward cue), at least for a certain proportion of
the trials. Indeed, the monkey had been extensively trained and knew the cues very well.
Additionally, the monkey could have relied on Pavlovian associations (Dickinson and Balleine,
1994). Such dual strategies of response could account for the bimodal distribution of action-
selection times (Figure 51A), with the fast times corresponding to trials in which the monkey
responded automatically, and slow trials corresponding to more complex computations and
investment of cognitive control. In coherence with that hypothesis, long action-selection times
were associated to lower sensitivity of choices to Forces, indicating that longer action-
selection times may have been associated to cognitive control exertion for the decision
process, facilitating decisions away from the default of choosing low-effort options (Epstein,
28 Note that it cannot be excluded that, in some situations, engaging in the action may have carried disutility (the
expected value of engaging may have been negative), in case of fluctuations in perceived reward values and effort costs (see Figure 11, Shenhav et al., 2013), and that refusals to engage (at least some of them) still allowed maximizing utility. Additionally, such disutility of engaging has been theorized in economics in the case of high-risk situations (Von Neumann and Morgenstern, 1947), and this effect could account for the refusals of high-force trials in the force-discounting tasks, in which high-Force category performance was relatively low compared to the performance in low-Force conditions.
Discussion
243
1994; Sloman, 1996). Note, still, that this assumption remains speculative, as the behavioral
analyses did not provide any additional evidence for the relationship between action-selection
times and cognitive control exertion for choices between offers.
2. Cognitive control for triggering actions
Once the monkeys had reached the decision to engage in courses of action involving
force exertion or delay to the Reward, cognitive control was probably needed to trigger the
action. Action-triggering can be thought of as a separate decision process, lasting from the go-
signal to the onset of the action. Indeed, in the one-option Force-discounting task used in
Study 1 (same dataset), Jahn and colleagues (2020) showed that fixation breaks signaling
disengagement from trials virtually always happened before go-signal onset, suggesting that
monkeys decided whether to engage between cue onset and go-signal onset, and only
triggered the action after the go-signal. We made the assumption that the same was true in
the Delay-discounting task, in which we could not precisely time refusals since gaze fixation
was not imposed. Additionally, in the two-alternative Force discounting task, gaze patterns (in
trials in which the monkey made a choice) were compatible, in most trials, with a sequential
process of choosing between alternatives before triggering the chosen action, a temporal
dissociation also described in EEG data (Dmochowski and Norcia, 2015).
In the context of value-based decision-making, RT had repeatedly been shown to scale
with decision difficulty (Hunt et al., 2012; Krajbich et al., 2010; Krajbich and Rangel, 2011;
Milosavljevic et al., 2010; Morris et al., 2019; Philiastides and Ratcliff, 2013), and to increase
with offer costs (Hunt et al., 2012; McGuire and Botvinick, 2010). In the two-alternative Force-
discounting tasks and the Delay-discounting task, the time taken to trigger the actions29
decreased with the reward and increased with the cost of the triggered option, as well as
during periods of time and in task conditions associated to low engagement rates. This
indicates that the time taken to trigger the actions reflected the difficulty of the decision
process of triggering the action. Note that we could not find any strong evidence for the
relationship between RT and the difficulty of triggering actions in the one-option Force
discounting task, supposedly because of strong motor effects on RT.
29 RT in the one-option tasks, middle press RT and action-triggering time in Study 2.
Discussion
244
Under the assumption that the rate of resource mobilization for cognitive processes is
constant across time, i.e. that longer time taken for any process is associated to increased
resource mobilization, the time taken to trigger the actions would have scaled with the
cognitive control invested in that decision process. Indeed, triggering the action was binary
(either successful or not, no in-between). Therefore, the optimal amount of control invested
in the action-triggering process would be exactly proportional to the difficulty, and therefore
just sufficient to trigger the action. This constitutes a rather speculative reasoning, which could
tend to be circular (Navon, 1984; Shenhav et al., 2017), thus the relationship between the time
taken to trigger actions and the cognitive control invested in the process remains hypothetical.
3. Physical effort to execute presses
In two of the force-discounting tasks used, reward delivery was conditioned by the
exertion of specific amounts of force on a grip. Following a long time taken to trigger the
actions, in the two Force-discounting tasks, monkeys usually exerted less force on the grip.
This means that if the monkey perceived force exertion as aversive, he took longer to trigger
the chosen action and did not press as hard on the grip. In other words, the monkey had a
stronger bias towards avoiding actions perceived as more costly (subjectively more difficult)
and was not as good at executing them i.e. at producing the physical effort required. This
pattern is compatible with the existence of a global computational variable, effort,
representing the resources required for specific courses of action, and encompassing both
their aversiveness when making choices, and therefore the difficulty to trigger them, and the
difficulty of executing them (Borderies et al., 2020). When the cost of specific actions
increased, the monkey was less willing to trigger them and had more difficulties mobilizing
the required resources to execute the action.
However, there could be at least one alternative interpretation for the negative
relationship between the time taken to trigger actions and the force exerted. Mental fatigue
caused by cognitive control exertion has a well-described negative impact on physical effort
performance (e.g. Boksem and Tops, 2008; Martin et al., 2018; McMorris, 2020; Schiphof-
Godart et al., 2018; Van Cutsem et al., 2017; reviewed by Habay et al., 2021). Under the
assumption that the time taken to trigger the action scaled with the cognitive control exerted
Discussion
245
to trigger it, such “ego-depletion” effects could have explained the decrease in force exerted
observed following long RTs.
B. LC activation may signal the end of decision
processes
247 LC neurons from five monkeys were recorded in four tasks: three effort-
discounting tasks (a Delay-discounting task and two Force discounting tasks, one-option and
two-alternative) and a target-detection task. In all three effort-discounting tasks, LC neurons
increased their rate both around cues onset (for one-option force-discounting task, see Jahn
et al., 2020 and Varazzani et al., 2015 for details on cue-related activity) and just before
action(s) onset (bar press(es) in Force-discounting tasks and bar release in Delay-discounting
task). This pattern is highly reminiscent of what was previously found in a variety of tasks (cue
onset: Bouret et al., 2012; Bouret and Richmond, 2009; Jahn et al., 2020; Varazzani et al., 2015;
action onset: Bouret et al., 2012; Bouret and Richmond, 2015, 2009; Bouret and Sara, 2004;
Clayton et al., 2004; Varazzani et al., 2015). The second study extended this finding to a task
with no artificial decomposition of the different steps of decisions, as monkeys were not
forced to wait between cues presentation and action onset, and could produce the chosen
actions as soon as they wanted. Identically, LC neurons displayed increased activity after cues
onset and before action onset, although the decreased temporal separation between these
events in the new paradigm made the distinction more difficult to assess.
LC activity has been proposed in the past to signal decisions (Clayton et al., 2004;
Rajkowski et al., 2004) and specifically their end (Shea-Brown et al., 2008). Moreover, pupil
diameter, a proxy for LC activity (Joshi et al., 2016; Varazzani et al., 2015), is known to increase
around the time of decisions (Beatty, 1982; de Gee et al., 2014; Gilzenrat et al., 2010; Lempert
et al., 2015; Nassar et al., 2012; Figure 6A), even in the absence of immediate responses
materializing them (Einhäuser et al., 2010). In value-based decision tasks, including the Force-
and Delay- discounting tasks presented earlier, the discharge of LC neurons could have been
evoked by the conclusion of two decision processes assumed to happen between cues onset
and press onset: choosing to engage (and between offers in case of multiple offers) after cues
onset, and triggering the (chosen) action. In the two-alternative choice task, the timing of the
end of the action-selection process was approximated using the last saccade onto chosen
Discussion
246
offer. However, I could not identify whether LC spikes occurring between cues and last
saccade onset, or between last saccade and side press onset, were better aligned onto either
one of these events. This negative result must be taken with caution, as it cannot be excluded
that it came from a lack of statistical power, or heterogeneity of the population of neurons.
Still, it could indicate that LC activity was neither evoked by cues onset, by the saccade onto
the chosen offer or by action onset, but by the termination of decision processes. Indeed, LC
neurons could, first, signal the end of the action-selection decisions. These decisions were
probably initiated by cues onset, and their end led to the saccade onto the chosen offer. Then,
before action onset, LC neurons could respond to the end of the action-triggering decisions.
These decision processes probably started around the time of the saccade and their end led
to action onset. The possibility that LC neurons may activate around the end of decisions had
already been suggested by Shea-Brown and colleagues (2008).
If that were true, first, it would explain why latencies of responses after cues onset in
the two-alternative Force-discounting task (Study 2, Figure 59A) seemed to be longer than
those previously observed in one-alternative tasks, in our first study and others (Figure 41A;
Fig. 3G from Varazzani et al., 2015, Fig. 3C from Bouret et al., 2012). Indeed, choosing whether
to engage and which offer to pick would take longer than deciding whether to engage when
facing only one offer, due to differences in the amount of information to process (Bettman et
al., 1990, 1990; Carpenter and Just, 1975; Huber, 1980; Johnson and Payne, 1985; Newell and
Simon, 1972; Payne et al., 1988; Shugan, 1980). Second, LC evoked responses by decision
termination could provide a mechanistic explanation for the generally higher variability in
latencies of cue-related responses compared to action-related responses (Delay-discounting
task: Figure 41A, and Fig. 3A from Bornert and Bouret, in prep; Fig. 3G&H from Varazzani et
al., 2015; Bouret et al., 2012, Fig 3C&D). Identically, the higher variability in the latencies of
pre-action responses in case of triggering complex motor actions compared to simple ones
could result from a higher variability in the time required to generate and transmit complex
motor commands at the end of the action-triggering process (differential force requirement
vs bar release, Bornert and Bouret, in prep, Fig. 3A&B). Furthermore, it could account for the
apparent absence of increase in LC activity after cue onset when cues were previously known
(Jahn et al., 2020), as the decision whether to engage may have been made even before cues
appeared, and may not be sufficiently aligned onto cue onset to detect coherent changes in
LC firing.
Discussion
247
In the target-detection task, LC neurons displayed one peak of activity, during the
action. In tasks involving perceptual decisions (visual or olfactory), LC neurons had repeatedly
been shown to respond only once during trials, with no response to stimulus (cue) onset and
only a response to action onset (Bouret and Sara, 2004; Clayton et al., 2004; Rajkowski et al.,
2004). In the theoretical frame explained above, it would mean that LC neurons only
responded at the end of the action-triggering process, but not to the target-detection process.
In other words, it would suggest that LC activity is in fact only evoked by the termination of
decisions associated to some cognitive control exertion. In that sense, the increase in LC
activity observed around cues onset in value-based decision task probably does not have the
same cause and function as the well-described peak of LC activity observed in response to
surprising stimuli such as sudden sounds (Bouret and Richmond, 2009; Bouret and Sara, 2004;
Grant et al., 1988; Vankov et al., 1995, Figure 20).
The pattern of activation of LC units observed in our target-detection task and other
perceptual tasks however contrasts with findings by Kalwani and colleagues (2014), who
showed that, in a Go/No-Go task, LC neurons responded both to the go signal and to action
onset (saccade). This difference in pattern may in fact reinforce the idea that LC neurons
respond at the end of costly decision processes. Indeed, in the countermanding task used by
Kalwani and colleagues (2014), when the Go-signal was detected, even in the absence of
doubt regarding its presence, there was an uncertainty regarding the correct response to
produce, as the stop-signal could have appeared later on during the trial. Consequently, the
monkeys may have exerted some cognitive control to deliberate and decide whether to
produce the saccade. In other words, in the task used by Kalwani and colleagues (2014), there
may have been an additional decision process to decide whether to engage based on the
probability of the Go-response being correct, compared to tasks used by Bouret and Sara
(2004), Clayton and colleagues (2004) and Rajkowski and colleagues (2014). Consequently, LC
neurons may have responded to the termination of that decision process. Additionally, LC
neurons have been hypothesized to encode uncertainty (Jepma et al., 2018; Jepma and
Nieuwenhuis, 2011; Nassar et al., 2012; Payzan-LeNestour et al., 2013), and this evoked
response to the go-signal in the task by Kalwani and colleagues may have reflected the
uncertainty about the correct response (Go or No-Go).
Discussion
248
Interestingly, in the task used by Kalwani and colleagues (2014) and our target-
detection task, the action-related increase in activity of LC neurons occurred after action
onset. This may have been the consequence of the implementation of the motor command
(action onset) being faster than the onset of the activation of LC neurons evoked by the
termination of the action-triggering process.
In the three tasks involving force exertion on a grip, LC neurons were active during the
press, and especially during the first half of the press. The timing of this activation was
compatible with it happening between press onset and peak of force attainment, as it started
after action onset and lasted about the same time as the onset-to peak time. This report of
increased activity during strenuous actions is coherent with previous reports of LC single unit
activity (Varazzani et al., 2015) and pupil dilation increases during force exertion (Varazzani et
al., 2015; Zénon et al., 2014). For one of the presses (middle press in two-alternative choice
task), I could determine that LC activity following press onset was in fact better aligned onto
the end of the ascending phase of the force exertion, i.e. the attainment of the peak in force.
This could indicate that LC neurons activated at the end of the contraction, i.e. signaled the
end of the motor program. For the other press (chosen action) of that task, I could not identify
whether LC activity was better aligned onto the onset of the press or the attainment of the
peak of force. This could have resulted from the increased diversity of movements required,
which could have introduced variability in the time taken to terminate the motor action after
the onset of the corresponding motor signal. Indeed, the side presses involved the use of two
grips and required different levels of force exertion.
The timing of responses of LC neurons to task events, probably at the very end of
decision processes, is quite incompatible with a direct influence on these decision processes.
For instance, consistently across our tasks and those used across numerous other studies, the
activation of LC neurons evoked by action onset appeared just prior to it (Bouret et al., 2012;
Bouret and Richmond, 2015, 2009; Bouret and Sara, 2004; Clayton et al., 2004; Varazzani et
al., 2015), or even sometimes just after it (Kalwani et al., 2014). Since LC axons are not
myelinated and conduct inputs rather slowly, pre-action LC activity would eventually reach
targets at least 130ms after their generation in LC neurons, i.e. too late to participate in the
decision process itself, especially if action-evoked activity did not precede the action. Bouret
Discussion
249
and Sara (Bouret and Sara, 2005; Sara and Bouret, 2012) have proposed that LC activity could
serve as a reset signal, facilitating or even causing the transition from one cognitive state to
another by provoking network reorganization. Although the evidence for an evoked response
of LC neurons to the end of decision processes (for instance selecting courses of actions and
triggering them) or motor processes (presses) is at this point mostly indirect and
circumstantial, it would be compatible with that theoretical framework. Still, additional
studies would be needed to provide more precise dissections of the timing of LC activity with
regards to decisions. Additionally, recording simultaneously in candidate target areas for
network reconfiguration and the LC could provide a direct confirmation of the network reset
theory.
C. LC neurons encode subjective difficulty, and
possibly effort, to engage in actions and execute them
Reliably across three tasks mobilizing cognitive effort to trigger costly actions, LC
neurons encoded the time taken to trigger the actions, before they were triggered, over and
above the encoding of task parameters influencing the difficulty of the action. As discussed
before, this time probably reflected the subjective difficulty to engage in actions, and may
have scaled with the cognitive control invested in triggering the action. Additionally, the fact
that the encoding of RT (or more specifically action-triggering time in Study 2) was present
even when including task parameters in the models indicates that this encoding was not
merely due to the correlation between RT (or action-triggering time) and task parameters, i.e.
objective difficulty. Rather, LC neurons encoded the subjective difficulty of triggering actions
on a trial-by-trial basis, and possibly the cognitive control invested. This encoding of the time
taken to trigger actions is reminiscent of positive correlations observed between RT and either
pupil diameter or population LC activity around action onset (respectively: de Gee et al., 2020;
Gilzenrat et al., 2010, experiment 3, and: Bari et al., 2020, Figure 25).
Although we could not provide any strong evidence for the consistency of this link
between RT and the subjective difficulty of triggering the press in the one-option Force
discounting task (Study 1), probably due to strong motor effects on RT, LC neurons still reliably
encoded RT prior to press onset. Conversely, we could not find any encoding of RT prior to
press onset in the target-detection task, in which RT was clearly affected by similar motor
Discussion
250
effects, but very probably not by subjective difficulty effects. Obviously, we cannot exclude
that the absence of reported effect of RT around press onset in the target-detection task did
not merely result from a lack of statistical power due to a low number of neurons. Still, we are
not the first to report an absence of relationship between LC activity and RT in conditions in
which RT does not reflect the cognitive control required to trigger the action. Indeed, Bari and
colleagues (2020) reported a positive correlation between RT and LC activity between target
onset and response onset only in the presence of distractors in a modified Posner task, but
not in the presence of neutral cues (Figure 25). This suggests that, in their task too, LC neurons
probably encoded RT only when it was related to decision difficulty, i.e. when distractor
presence called for cognitive control investment to produce the correct response, and not to
sensory-motor demands, i.e. in the absence of a distractor.
In the two tasks mobilizing physical effort through differential force exertion on a grip,
LC neurons additionally encoded the required force and the force exerted on the grip during
the presses, suggesting that they encoded the difficulty of the action and probably the physical
effort associated. Interestingly, this effect was not present for the small force exertion
required in the Target-detection task or the middle press of the two-alternative Force-
discounting task. One possible explanation for these results would be that the required force
was so small that the monkey did not perceive variations of exerted force as variations in effort
costs. Note that in the two Force-discounting tasks, around the time of the press requiring
differential force exertion, we could not find any encoding of the exerted force on the grip by
LC neurons if we restricted our analysis to trials in which the small Force was executed.
Similarly, in a Force-discounting task with four Force requirement levels, Zénon and colleagues
(2014) did not report any effect of the increase in exerted force between the two conditions
with the smallest Force requirements on pupil dilation during the press (Figure 3A). These
findings argue for an encoding of the subjective difficulty of the press rather than of the force
actually exerted on the grip.
Overall, the findings reported throughout the two studies indicate that LC neurons
encoded the subjective difficulty, and possibly the invested cognitive and physical effort of
triggering and executing actions. In other words, LC neurons probably encoded the effort
produced at the time of its production, and this encoding was domain-general: both physical
Discussion
251
and mental effort were encoded. The question of the source of these modulations cannot be
addressed with certainty by the results presented, but a hypothesis can be formulated. A vast
body of literature reports a global sympathetic activation, evidenced by pupil diameter
increases, in response to both physical (Varazzani et al., 2015; Zénon et al., 2014) and mental
effort exertion (Alnæs et al., 2014; Hess and Polt, 1964; Kahneman and Beatty, 1966; Simpson
and Hale, 1969; Wahn et al., 2016; Wright et al., 2013). The encoding of subjective difficulty,
and possibly exerted effort, by LC neurons could result from inputs from the autonomic
system, in particular the nucleus Paragigantocellularis (PGi). The PGi is a nucleus from the
rostral medulla responsible for the respiratory and cardiovascular adaptations triggered by
sympathetic activation, and one of the two major inputs to the LC (Aston-Jones et al., 1991,
1986; Ennis and Aston-Jones, 1988, 1986; Van Bockstaele and Aston-Jones, 1995). However,
Ennis and Aston Jones (1988) report very short latencies between PGi activation and LC
activation, as inputs from the PGi to the LC produced LC responses around 11ms after their
onset. Autonomic activation increases during effort production. Under the assumption that
the PGi does encode an approximation of the mental or physical effort produced, via an
encoding of the degree of autonomic activation, PGi activity would be higher throughout
periods of effort production, such as decision periods. Given the fast effects of PGi activity on
LC activity, one would expect effort production-associated LC responses to appear from the
beginning of effort production. In our tasks, decisions provide an example of processes
associated to effort production that last for a few hundred milliseconds. Contrary to what
would be expected if effort-related modulations of LC activity were driven by PGi inputs, we
showed that there is converging evidence that the LC activates and encodes the effort
produced for the decisions at their end.
The modulations of LC activity with subjective difficulty (and possibly effort exertion)
could also result from inputs from cortical regions known to encode effort: the dorsal anterior
cingulate cortex (dACC), the anterior insula and the posterior parietal cortex (e.g. Burke et al.,
2013; Bush et al., 2002; Croxson et al., 2009; Hare et al., 2011; Kurniawan et al., 2013; Prévost
et al., 2010; Samanez-Larkin et al., 2008; Skvortsova et al., 2014). The ACC is of particular
interest as fMRI studies (Croxson et al., 2009; Forstmann et al., 2006; Mars et al., 2005; Morris
et al., 2019; Shenhav et al., 2014; Walton et al., 2004; Yoshida and Ishii, 2006) as well as single-
unit recordings (Hillman and Bilkey, 2010; Kennerley et al., 2009, Lespart and Bouret,
unpublished data) have reported increased activity during decisions as well as difficulty and
Discussion
252
effort encodings, in both physical effort tasks and cognitive control tasks (at least decision-
making tasks). Considering the very variable definitions of brain areas across studies, refining
the location of those encodings and dissecting the projections of the cortical areas encoding
effort onto the LC could be critical to better understand the circuitry underlying effort-related
LC activity. Additionally, considering the large number of inputs received by each LC unit
(Schwarz et al., 2015), LC units could in fact multiplex signals from various brain areas and
combine them into a subjective difficulty/effort signal.
D. LC activity and value-based decision-making
In three tasks involving cost-benefit tradeoffs, Force- or Delay-discounting tasks, LC
neurons displayed increased activity after cue(s) onset. Consistently across tasks, and
coherently with previous single-unit recording studies of LC activity (Bouret et al., 2012;
Bouret and Richmond, 2015; Jahn et al., 2020; Varazzani et al., 2015), LC neurons were found
to encode Reward(s) in both Force- and Delay-discounting tasks. In addition to that, only in
the Delay-discounting task were LC neurons found to encode the cost of the trial, namely
Delay, after cues onset. In contrast, in Force-discounting tasks (Jahn et al., 2020; Varazzani et
al., 2015) or reward schedule tasks (Bouret et al., 2012) the cost (Force or sequence length,
respectively) was not encoded. This pattern of strong encoding of Reward-related information
associated to faint or absent encoding of effort-related parameters during decisions is
reminiscent of the pattern of activity observed in the vmPFC (Kable and Glimcher, 2007; Peters
and Büchel, 2009; Plassmann et al., 2010; Prévost et al., 2010; San-Galli et al., 2018; Skvortsova
et al., 2014; Talmi et al., 2009), suggesting that the responses of LC neurons to cues onset may
result from an interplay with the vmPFC. It is unclear, however, whether the vmPFC would
feed the LC with reward-related information or whether it is the other way round. San-Galli
and colleagues (2018), report very short latencies to Reward encoding in the vmPFC (less than
200ms, San-Galli et al., 2018, Fig. 4A&D), quite similar to those reported in the same task by
Varazzani and colleagues (2015, Fig. 5C). Considering the slow transmission of LC inputs to the
cortex (above 100ms, Aston-Jones et al., 1985), it would seem more probable that LC neurons
receive inputs from the vmPFC or from the same area that feeds reward information to the
vmPFC, rather than the opposite.
Discussion
253
Engaging in actions leading to delayed Rewards is costly, as it requires engaging
cognitive control (Aranovich et al., 2016; Hirsh et al., 2008; Kurth-Nelson et al., 2012; Sasse et
al., 2017; Shamosh et al., 2008), and in our Delay-discounting task, monkeys did discount
Rewards with Delay, as they exhibited lower engagement rates for high-Delay offers.
Additionally, the marginal encoding of condition-by-condition engagement rate (over and
above Reward and Delay), and the coherent strength of encoding of Reward and Delay across
neurons constitute two arguments towards a representation of the discounted value of the
offered reward by LC neurons around cue onset, in the particular case of delay-discounting.
These results are in contrast to previous LC single-unit recordings in Reward-schedule tasks.
In such tasks, sequence length, which produces a delay to the reward, has been shown to
produce, in a majority of subjects, similar effects on engagement rates as delay to the reward
(Minamimoto et al., 2012), yet it is not encoded by LC neurons (Bouret et al., 2012). Such
contrast with our Delay-discounting task could come from differences in task structure.
Indeed, in our Delay-discounting task, engaging in the action (releasing the bar) predicted
Reward delivery with certainty. Conversely, in Reward-schedule tasks, repeated engagement
for each press of the sequence was required to ensure reward delivery, and each cue could in
theory lead to a decision whether to engage.
Additionally, pharmacological decreases of LC activity using α2 agonist administration
(clonidine) in monkeys performing a Reward-schedule task have failed to provoke any change
in the modulation of engagement probability by offer parameters, or any change in
engagement rates in sequences or individual presses (Jahn et al., 2018). This result argues for
an absence of involvement of LC activity in engagement in planned sequences of actions based
on the discounted value of the reward, and in individual decisions to engage within sequences
of actions. In other words, the activity of LC neurons after cue onset reflects the reward and
its characteristics: size, delay to delivery, and possibly uncertainty (Jepma et al., 2018; Jepma
and Nieuwenhuis, 2011; Nassar et al., 2012; Payzan-LeNestour et al., 2013). However, it is not
likely to directly affect decisions to engage, and in particular could not mediate the effect of
Reward or Delay (and possibly uncertainty) on engagement, at least in case of well-known
cues, i.e. outside of learning conditions.
In Borderies et al. (2020), in a two-alternative Force-discounting task, we reported that
clonidine administration, i.e. decreases in LC activity, selectively affected the probability of
Discussion
254
choosing the high-Effort options in two ways: a global decrease in the overall tendency to
choose the high-effort offers and an increase in the sensitivity of choices to the Effort
dimension. In a similar task (Study 2), I report no encoding of the effort dimension of the
offers, only of the Reward dimensions, after cues onset. Although we cannot exclude that this
result comes from a lack of statistical power, or from the lower sensitivity of the monkey’s
choices to Forces than to Rewards, these results tend to indicate that LC neurons cannot
participate in the effort-based valuation of offers, even in the context of goal-directed action-
selection. Conversely, I proposed an indirect mechanism for the relation between LC activity
and choices of high-Effort courses of action via modulations of Effort sensitivity. Within trials,
I showed that increased LC activity around early steps led to decreased subjective difficulty in
the next step of the trial. Indeed, increased activity around the first press led to increased
probability to engage in the second press, and increased activity around cues onset led to
decreased time taken to trigger the second press. In other words, increased activity early in
trials was associated to decreased subjective difficulty to engage in costly actions. Evidence
for a causal link between LC activity and the subjective difficulty to engage was provided by
Jahn and colleagues (2018), who showed that clonidine administration led to increased RTs,
which probably reflected the subjective difficulty to engage in actions.
Additionally, I showed that increased activity around cues onset was associated to
increased probability of choosing the high-Effort offer, and to an increase in the exerted force
for the second press, especially in the most difficult conditions. This indicates that increased
early LC activity probably led to decreased subjective difficulty of executing the actions.
Decreases in LC activity with clonidine have been shown to cause a decrease in exerted force
(Borderies et al., 2020; Jahn et al., 2018). We excluded that this effect was a mere muscular
effect, i.e. a consequence of a decreased motor capacity to exert force (Borderies et al., 2020).
Consequently, this result provides evidence that the relation between LC activity and the
subjective difficulty to execute effortful actions was probably causal. Additionally, in Borderies
et al. (2020) we report that the decrease in exerted force and increased force sensitivity of
choices could be accounted for by a single effort sensitivity variable, increased in case of
clonidine administration. The second study proposed a mechanism for this impact of LC
activity on value-based decision-making, and in particular effort-based decision-making.
In Study 2, I showed that increased LC activity around early steps of the trials requiring
effort production facilitated effort production later on. Importantly, this effect was domain-
Discussion
255
general, meaning that it LC activity facilitated physical and cognitive effort production later
on. Indeed, increased LC activity around the middle press was associated to increased
probability to engage in choices and faster chosen action-triggering, and increased activity
around cues onset led to decreased decision biases towards choosing easy offers. This
suggests that increased LC activity early in trials facilitated cognitive control exertion later on,
in order to override default courses of actions (refusing trials and choosing low-force options).
Moreover, increased cues-onset related activity led to increased force exertion during chosen
action execution, indicating that LC activity a little less than one second prior to action onset
may have facilitated physical effort production. In other words, our results were compatible
with domain-general delayed effects of LC activity on effort production, i.e. facilitation of
effort production regardless of the type of challenge faced (physical or cognitive). In Borderies
et al. (2020), long-term decreases in LC activity would have led to a decrease in the capacity
to produce effort, and therefore to an increase in the perception of effort. Consequently, this
would have led to an adjustment of choices to avoid high-effort offers more often, especially
if a much less effortful alternative was available. This adjustment of choices would have had
to be within very few trials, as we reported no tendency for a drift in high-Effort choice
probability throughout sessions.
The existence of a monitoring unit has been theorized for control allocation (Shenhav
et al., 2013), and could be extended to physical effort exertion, especially if cognitive and
physical effort draw on similar resources (see part I.A.3). The monitoring unit collects
information on previous resource expenses in order to adjust the parameters of the future
decisions to allocate effort made by the selection unit (see part I.A.2.c). My data as well as
previously obtained results are compatible with a role of the LC as a provider of feedback to
the monitoring unit. Consistently across three effort-discounting tasks, LC neurons displayed
an encoding of the subjective difficulty and probably the exerted effort. I additionally showed
in Study 2 that higher activity around second press execution was associated to decreased
probability of choosing the high-Effort offer in the next trial. If LC neurons do encode the
exerted effort for the press, my results indicate that this signal could participate in the
adjustment of future decisions to allocate effort. In addition to feedback regarding effort
production itself, LC neurons seem to provide feedback on some of its consequences
(outcomes). Jahn and colleagues (2020) reported that LC neurons respond to commitment
Discussion
256
errors (failures to engage in trials), and I complemented this finding by showing that LC
neurons respond to execution errors. Both these events can be used as signals of insufficient
control allocation. Botvinick and colleagues (Botvinick, 2007; Botvinick et al., 2004, 2001)
additionally theorized that response conflict detection could constitute another feedback
signal regarding the adequacy of control allocation with respect to task demands. Increased
LC activity (by tVNS) has been shown to facilitate behavioral adjustments to conflict (slowing
of RT), suggesting that LC activity could indeed provide a conflict detection signal provoking
an adjustment of control allocation (Fischer et al., 2018). NA has been repeatedly shown to
have a causal role in adaptations to changes in task rules (Lapiz et al., 2007; Lapiz and Morilak,
2006), especially via LC projections to the cortex (McGaughy et al., 2008; Tait et al., 2007).
Importantly, higher NA levels (higher atomoxetine doses) are required to facilitate set shifting
than to enhance working memory and inhibitory control (Newman et al., 2008). Error-related
LC activity could participate in set-shifting, by facilitating adjustments to new contingencies.
The targets of LC activity that account for its delayed effects on effort production
cannot be determined by the studies presented in this manuscript, but suggestions, opening
the path for further research, can be formulated. The ACC has strong reciprocal connections
to the LC (Schwarz et al., 2015). It has been shown to multiplex Reward- and Effort-related
information (Amiez et al., 2006; Hayden and Platt, 2010; Kennerley et al., 2011) and to have a
causal role in effort-based decision-making (Rudebeck et al., 2006; Walton et al., 2003, 2002).
Specifically, the ACC has been implicated in learning contingencies and consequences of effort
exertion (Rudebeck et al., 2006; Rushworth et al., 2004; Walton et al., 2009, 2004, 2003), such
as rewards and errors (Fiehler et al., 2004; Jahn et al., 2014; Quilodran et al., 2008; Ullsperger
and Cramon, 2003). Interestingly, error-related responses of the ACC are reminiscent of those
observed in the LC by Jahn and colleagues (2020) and myself in Study 2, since ACC neurons
respond to task disengagement, more strongly for bigger offered rewards (Amiez et al., 2005).
More generally, the ACC is believed to be critical for updating internal models of the
environment (Kolling et al., 2016; O’Reilly et al., 2013), planning sequences of actions (Abe
and Lee, 2011; Barraclough et al., 2004; Tervo et al., 2014; Zhu et al., 2012), and triggering
behavioral change (Ridderinkhof et al., 2004; Rushworth et al., 2004). These elements have
led several teams to propose a role of the ACC as the monitoring unit of control, collecting
feedback on effort production and its consequences to adjust and plan future courses of
Discussion
257
actions, and possibly as the specification unit as well (Botvinick, 2007; Botvinick et al., 2004,
2001; Shenhav et al., 2013).
Moreover, the role of the ACC on high-effort choice promotion has been shown to
depend critically on catecholamine neurotransmission (Schweimer et al., 2005), and more
precisely DA fixation on D2 receptors (Walton et al., 2009), but not on nucleus accumbens
integrity (NAc, Walton et al., 2009, 2005). Several studies have demonstrated that
modifications of LC activity lead to modifications of DA levels in cerebral cortices, including
the ACC (Devoto et al., 2003, 2001; Kawahara et al., 2001), probably due to a co-release of DA
with NA (Devoto et al., 2005). These results indirectly suggest that LC activity could participate
in the putative role of the ACC in monitoring effort exertion, possibly by providing it with
feedback on effort production and errors. Thus, it would be interesting for further studies to
explore this hypothesis. Additionally, certain ACC neurons exhibit plasticity (pyramidal
neurons, Lee et al., 2007; Liu et al., 2018; Shyu and Vogt, 2009; Toyoda et al., 2011), and NA
has plasticity-promoting effects in some brain regions (Bing et al., 1992, 1991; Gubits et al.,
1989; Maity et al., 2020; Sonneborn and Greene, 2021; Stone et al., 1997, 1991; Walling and
Harley, 2004). If NA does promote plasticity in the ACC, it could participate in the adjustments
of models of the environment and of effort allocation strategies, in response to past effort
production.
E. LC and facilitation of effort exertion: mechanistic
hypotheses
Previous studies exploring the link between LC activity and resource mobilization for
cognitive and physical processes had shown a causal role of LC activity in physical effort
production (Borderies et al., 2020; Jahn et al., 2018). In Study 1, we showed that LC neurons
encoded the subjective and objective cost of triggering and executing actions, at the time of
action triggering and execution, respectively. Considering the slow conduction velocity of LC
axons (Aston-Jones et al., 1985), the timing of such encodings was only compatible with effects
on ulterior processes, which could hardly be tested in single-step one-option tasks. In the
second study, using a two-step two-alternative force-discounting task, I showed that, within
trials, LC activity had a negative effect on the subjective difficulty of triggering and executing
actions, suggesting that it may have energized behavior, such that triggering and executing
Discussion
258
the actions was consequently perceived as less difficult. Importantly, the time elapsed
between the LC activity modulating a certain behavior and the behavior itself was rather long,
around one second, suggesting either very delayed or quite long-lasting effects, or both. Such
delayed effect of LC activity on behavior had previously been noted by Bari and colleagues
(2020), who reported that increased baseline activity of LC neurons, i.e. before distractor
onset, in a modified Posner task, was associated to lesser effects of distractors on behavior
and better performance. In other words, increased LC activity facilitated inhibitory control, a
subcategory of cognitive control. In a way, such effects of pre-trial onset activity on
subsequent behavior are reminiscent of the relation between LC activity and subjective
difficulty in an epoch starting before cues onset, i.e. anticipating the cues.
Effort exertion is believed to be aversive due to the consumption of some resources,
although the exact nature of such resources remains debated (Shenhav et al., 2017).
Consequently, increasing the amount or biasing the perception of the resources available
would logically make subjective difficulty decrease. Astrocytic glycogen as well as brain
glucose have been proposed as the limited quantity that the effort allocation system aims at
sparing when allocating cognitive control with caution (Baumeister et al., 1998; Baumeister
and Heatherton, 1996; Christie and Schrater, 2015; Muraven et al., 1998). As for physical
effort, the brain energy demand increases during physical exercise (Secher et al., 2008),
causing an increase in lactate and glucose consumption (Larsen et al., 2008; Vissing et al.,
1996). This increase is met by higher glycogen degradation in multiple brain areas, in particular
but not restricted to brain areas involved in motor control, an effect that could not be solely
accounted for by the muscle-activity induced decrease in blood glucose (Matsui et al., 2012,
2011). Additionally, this depletion in glycogen has been linked to endurance capacity and
fatigue following prolonged physical effort (Matsui et al., 2019, 2017), suggesting a role of
glycogen reserves in physical effort perception i.e. subjective difficulty.
LC activity and NA release are associated to modifications of astrocytic metabolism
(Cohen et al., 1997; Gibbs, 2015; Gibbs et al., 2010; Hertz et al., 2010; Müller, 2014; O’Dowd
et al., 1995; Quach et al., 1978). First, NA induces short term glycogen degradation into lactate,
an immediately available substrate for neuronal metabolism (Magistretti, 1988; Quach et al.,
1978; Sorg and Magistretti, 1991). In the longer term, NA promotes glycogen synthesis via
transcriptional effects (Allaman et al., 2000; Sorg and Magistretti, 1992). Chemical reactions
Discussion
259
(including those happening in living cells) follow an equilibrium (Lodish et al., 2000), meaning
that the final concentrations of the product depend on the initial concentration of the
substrate. Consequently, if glycogen levels decrease, the amount of lactate that can be
released decreases. In the short term, NA appears to bias the reaction transforming glycogen
into lactate, possibly by modifying enzymatic activity within astrocytes (Uhr, 1979) via
increases in cAMP (Fink et al., 2021), and could therefore be making lactate more available for
brain areas involved in the task at hand regardless of glycogen supplies. The increase in lactate
availability following NA release has been shown to be beneficial for some cognitive processes
such as memory formation and consolidation (Gibbs, 2015; Gibbs et al., 2010). Although, to
my knowledge, the direct link between NA release in the brain and cognitive control or
physical effort performance via increases in available metabolic substrate has not been
directly demonstrated, it can he hypothesized that the effects observed for memory could
generalize to more (if not all) brain areas, and therefore other processes. Additionally, the
dynamics of NA-dependent increases in astrocytic and neuronal lactate have been examined,
in the context of arousal-induced NA release. Lactate increases have been shown to have a
relatively delayed onset (a few seconds, Zuend et al., 2020), but additional studies with
increased temporal precision and, importantly, in the context of task execution and effort
production, would be required to further demonstrate the possible relation between
metabolic adjustments and effort facilitation by LC activity.
The aversiveness of control exertion has also been proposed to be caused by the
existence of bottlenecks in processing systems generating a risk of interferences between
cognitive operations (Allport, 1954; Logan, 1985). NA release is known to affect brain
functional connectivity (Coull et al., 1999; Guedj et al., 2017b; Hernaus et al., 2017; Zerbi et
al., 2019). Pharmacological modifications of brain NA levels during task execution have been
shown to coherently modulate performance and the connectivity between task-related areas
(atomoxetine: Brink et al., 2016; Hernaus et al., 2017; clonidine: Coull et al., 1999), suggesting
a causal role for NA-mediated modifications of brain connectivity in mental effort exertion.
Mather and colleagues (2016) have suggested that NA release in the presence of glutamate
could both enhance local activity and lateral inhibition, thus potentially decreasing the risk for
competition between neuronal networks. Glutamate, a major excitatory neurotransmitter, is
critical for cognitive function and especially cognitive control exertion (Dalton et al., 2011; Jett
Discussion
260
et al., 2017; Naaijen et al., 2018; Nardecchia et al., 2018; Stefani et al., 2003; Stefani and
Moghaddam, 2005; Yuen et al., 2012, 2011). Moreover, during cognitive control exertion,
intensity of neuronal activity (measured by fMRI) in critical regions such as the dorsal ACC, the
orbitofrontal cortex and the basal ganglia correlates positively with glutamate levels in these
areas (Falkenberg et al., 2012). In other words, glutamate is released in highly active areas
during control-greedy tasks, and this release is critical for cognitive control exertion. The
supposed synergy between NA and glutamate could therefore constitute a mechanism for the
facilitation of control exertion by LC activation.
F. Relationship between noradrenergic facilitation of
effort and promotion of behavioral variability
I proposed that the LC could be facilitating effort production in the short term while
providing feedback regarding current effort production to adjust future decisions. Multiple
studies have shown a causal role for the LC/NA system in behavioral variability and especially
decision noise (Dubois et al., 2021; Jahn et al., 2020; Kane et al., 2017; Tervo et al., 2014). Even
though this role may seem disconnected from effort facilitation at first glance, it may in fact
be a collateral effect of it. If the LC does provide feedback on effort production, the intrinsic
variability in LC activity across trials may cause trial-by-trial adjustments in effort allocation,
and therefore introduce decision noise. Additionally, since we showed that LC activity
influences trial-by-trial probability to exert control to engage in high-effort offers, trial-by-trial
variability in LC responses could account for some of the decision noise. Decreases in LC
activity (Borderies et al., 2020; Jahn et al., 2018) would lead to decreased feedback strength
about effort production, causing smaller trial-by-trial adjustments of effort allocation across
trials, and decreased variations in the enhancement of high-effort decisions. On average, this
would increase choice consistency. Noise in the neuronal activity has been proposed to cause
the noise in decisions (Drugowitsch et al., 2016; Faisal et al., 2008; Renart and Machens, 2014;
Wang, 2002), and I showed that the intrinsic noise in the LC neuronal activity could account
for (some of) decision variability, by injecting noise into decisions, as suggested by Doya
(2002). The proposal from Aston-Jones and Cohen (2005) that LC activity influences the noise
in the neuronal activity of target areas is not incompatible, but complementary.
Discussion
261
Tervo and colleagues (2014) have reported that stimulating LC inputs to the ACC
increased behavioral variability (noise), suggesting that the effects of LC activity on decision
noise could rely on LC-ACC projections. Interestingly, I proposed that effort-related effects of
LC activity would critically depend on LC-ACC projections as well. Still, it is unclear whether
artificial electrical LC stimulation properly mimics the few spikes evoked by task events,
justifying the need for additional studies of the effects of modifying the strength of LC inputs
to the ACC.
A second aspect of behavioral flexibility is the balance between exploratory and
exploitative behavior. Pupillometry studies have implicated LC activity in switches from
exploitation to exploration, suggesting that brief increases in LC activity (associated to
transient pupil dilations) could lead to task disengagement (Jepma et al., 2011), and that
periods of exploration could be associated to higher LC activity (associated to higher baseline
pupil diameter, Gilzenrat et al., 2010). The causal role for LC activity in task disengagement
was later demonstrated using DREADDs (Kane et al., 2017). The results described in this
manuscript, associated to previous studies, indicated that LC neurons probably signal the
subjective difficulty of cognitive and physical processes, and possibly the invested effort, and
that this signal serves to adjust future effort allocation. Task disengagement could constitute
an extreme adjustment in case of detection of very high subjective difficulty, i.e. following
strong LC activation.
G. Modularity of the LC: evidence for differences in
electrophysiological properties and consequences for
effort processing
Recent anatomo-functional studies of the LC have emphasized the heterogeneity of LC
organization, both regarding cell morphology (Loughlin et al., 1986; Swanson, 1976),
projection patterns (Schwarz et al., 2015), neurotransmitter synthesis (Sutin and Jacobowitz,
1991), electrophysiological properties (Totah et al., 2018; Williams et al., 1984) and effects on
cognitive processes (Uematsu et al., 2017). In the first study, neuron-by-neuron baseline
remained stable across the duration of the experiments (Bouret and Richmond, 2015, 2009;
Bouret and Sara, 2004; Clayton et al., 2004; Kalwani et al., 2014). However, I found that LC
Discussion
262
neurons exhibited variable baseline activity, correlating positively with the amplitude of
evoked responses to task events and their signal-to-noise ratio, and with the strength of
encoding of task and behavioral parameters, suggesting overall differences in excitability
across neurons. These findings reinforce the idea of the diversity in electrophysiological
properties across LC neurons, although the spatial precision of the recording method used
here did not allow identifying whether neurons with different properties were spatially
dissociable. Additionally, we had no way of assessing the projection pattern of the recorded
neurons, although rodent studies have suggested that excitability differences may be
associated to projection pattern heterogeneities (Chandler et al., 2014).
In Study 2, I reported that increased LC activity was associated to later decreases in
subjective difficulty perception both for cognitive and physical effort, in coherence with
pharmacological studies (Borderies et al., 2020; Jahn et al., 2018). Additionally, I provided
evidence for a potential role of LC activity in adjusting future control allocation. However, the
method of analysis considered LC neurons as a population, and it cannot be excluded that
different effects of LC activity resulted from different modules, with different projection
patterns. The detailed inspection of effects of pharmacological manipulations of LC activity by
clonidine administration (Borderies et al., 2020; Jahn et al., 2018) provides some insight into
this possibility. Indeed, α2 receptors are not evenly distributed throughout the LC, and are
more numerous in the posterior part (Chamba et al., 1991), a portion of the LC that
preferentially projects onto the thalamus (Schwarz et al., 2015), an area critically involved in
motor control (Sommer, 2003). Consequently, clonidine at lower doses would have
preferentially decreased posterior LC activity and therefore NA release into the thalamus.
Jahn and colleagues (2018) used several doses of clonidine (2.5 or 5 or 7.5μg/kg), all
lower than the dose we used by Borderies et al. (2020, 10 μg/kg). Motor effects (decreased
exerted force) were observed in both studies, and appeared even for the lowest dose of
clonidine (Figure 28B&C). These effects may in fact be dependent on projections of the
posterior part of the LC onto the thalamus. As detailed earlier, they may account for the
increase in sensitivity of choices to the effort dimension, although in Borderies et al. (2020)
we only used a high dose of clonidine, which probably had effects on the activity of other parts
of the LC. Conversely, Jahn and colleagues only reported effects on RT, and therefore possibly
cognitive effort, for the highest dose of clonidine used, suggesting that it may rely on separate
Discussion
263
LC neurons, in a more anterior part of the LC with lower α2 receptor concentration and
probably projecting onto separate brain regions (possibly frontal cortices).
H. Conclusions and perspectives
Altogether, my work provides some mechanistic insight into the involvement of LC
activity in motivational processes, and more specifically effort perception, production and
regulation of allocation. Throughout three effort-discounting tasks, LC neurons displayed
evoked responses with timings compatible with signaling the end of decision processes, and
encoded the subjective difficulty of these decisions and the cognitive control exerted. During
actions, LC neurons encoded their subjective difficulty and the physical effort involved. LC
activation was associated to modifications of subjective difficulty, and these modifications
appeared to be delayed (around 1s after the activation of neurons) and/or quite long-lasting.
Within trials, LC activity biased choices towards high-effort offers and facilitated action
execution, and I suggested that these effects may be due to metabolic and/or connectivity
effects of NA on targets. Across trials, LC activation reflecting the exerted effort and signaling
errors probably served as a learning signal to adjust future effort allocation. This effect would
be compatible with neuronal interplay between the LC and the ACC, the putative structure
monitoring effort production, with feedback on effort-related processes. Such meta-learning
role of the interplay between the LC and the ACC, in complement with an interplay between
the ACC and DA nuclei, had already been proposed in the context of (but not restricted to)
effort-discounting tasks (Silvetti et al., 2018). Consequently, my work clarifies the dynamics of
the implication of the LC in decision-making and supports a key role in effort perception,
production, and allocation adjustments.
Considering the anatomic and functional heterogeneity of the LC, a critical next step in
understanding the implication of LC activity in effort processing would consist in dissecting the
projection patterns of the LC in order to identify their relative contribution. DREADDs
associated with retrograde viruses now allow selectively inhibiting certain subpopulations of
neurons based on their projection patterns. Selectively inhibiting LC-ACC projections could
provide confirmation of the scenario proposed above regarding the role of the LC as an effort
feedback provider for the ACC, while modifications LC-thalamus projections may confirm the
contribution of these projections in physical effort facilitation.
References
264
References
Aarts, E., Roelofs, A., Franke, B., Rijpkema, M., Fernández, G., Helmich, R.C., Cools, R., 2010. Striatal Dopamine Mediates the Interface between Motivational and Cognitive Control in Humans: Evidence from Genetic Imaging. Neuropsychopharmacology 35, 1943–1951. https://doi.org/10.1038/npp.2010.68
Aasman, J., Mulder, G., Mulder, L.J.M., 1987. Operator Effort and the Measurement of Heart-Rate Variability. Hum. Factors J. Hum. Factors Ergon. Soc. 29, 161–170. https://doi.org/10.1177/001872088702900204
Abe, H., Lee, D., 2011. Distributed Coding of Actual and Hypothetical Outcomes in the Orbital and Dorsolateral Prefrontal Cortex. Neuron 70, 731–741. https://doi.org/10.1016/j.neuron.2011.03.026
Adams, L.M., Foote, S.L., 1988. Effects of locally infused pharmacological agents on spontaneous and sensory-evoked activity of locus coeruleus neurons. Brain Res. Bull. 21, 395–400. https://doi.org/10.1016/0361-9230(88)90151-7
Agster, K.L., Mejias-Aponte, C.A., Clark, B.D., Waterhouse, B.D., 2013. Evidence for a regional specificity in the density and distribution of noradrenergic varicosities in rat cortex. J. Comp. Neurol. 521, 2195–2207. https://doi.org/10.1002/cne.23270
Allaman, I., Pellerin, L., Magistretti, P.J., 2000. Protein targeting to glycogen mRNA expression is stimulated by noradrenaline in mouse cortical astrocytes. Glia 30, 382–391.
Allport, G.W., 1954. The nature of prejudice, The nature of prejudice. Addison-Wesley, Oxford, England.
Alnæs, D., Sneve, M.H., Espeseth, T., Endestad, T., van de Pavert, S.H.P., Laeng, B., 2014. Pupil size signals mental effort deployed during multiple object tracking and predicts brain activity in the dorsal attention network and the locus coeruleus. J. Vis. 14. https://doi.org/10.1167/14.4.1
Alvarez, V.A., Chow, C.C., Bockstaele, E.J.V., Williams, J.T., 2002. Frequency-dependent synchrony in locus ceruleus: Role of electrotonic
coupling. Proc. Natl. Acad. Sci. 99, 4032–4036. https://doi.org/10.1073/pnas.062716299
Alvarez-Maubecin, V., García-Hernández, F., Williams, J.T., Bockstaele, E.J.V., 2000. Functional Coupling between Neurons and Glia. J. Neurosci. 20, 4091–4098. https://doi.org/10.1523/JNEUROSCI.20-11-04091.2000
American Psychiatric Association, American Psychiatric Association (Eds.), 2013. Diagnostic and statistical manual of mental disorders: DSM-5, 5th ed. ed. American Psychiatric Association, Washington, D.C.
Amiez, C., Joseph, J.P., Procyk, E., 2006. Reward Encoding in the Monkey Anterior Cingulate Cortex. Cereb. Cortex 16, 1040–1055. https://doi.org/10.1093/cercor/bhj046
Amiez, C., Joseph, J.-P., Procyk, E., 2005. Anterior cingulate error-related activity is modulated by predicted reward. Eur. J. Neurosci. 21, 3447–3452. https://doi.org/10.1111/j.1460-9568.2005.04170.x
Anderson, J.R., 1990. The adaptive character of thought, The adaptive character of thought. Lawrence Erlbaum Associates, Inc, Hillsdale, NJ, US.
Anderson, J.R., 1983. The architecture of cognition, Cognitive science series. Harvard University Press, Cambridge, Mass.
Andrade, L.S., Kanitz, A.C., Häfele, M.S., Schaun, G.Z., Pinto, S.S., Alberton, C.L., 2020. Relationship between Oxygen Uptake, Heart Rate, and Perceived Effort in an Aquatic Incremental Test in Older Women. Int. J. Environ. Res. Public. Health 17. https://doi.org/10.3390/ijerph17228324
Aranovich, G.J., McClure, S.M., Fryer, S., Mathalon, D.H., 2016. The effect of cognitive challenge on delay discounting. NeuroImage 124, 733–739. https://doi.org/10.1016/j.neuroimage.2015.09.027
Ariely, D., Gneezy, U., Loewenstein, G., Mazar, N., 2009. Large Stakes and Big Mistakes. Rev. Econ. Stud. 76, 451–469.
References
265
Arnsten, A.F., 1997. Catecholamine regulation of the prefrontal cortex. J. Psychopharmacol. Oxf. Engl. 11, 151–162. https://doi.org/10.1177/026988119701100208
Arnsten, A.F., Cai, J.X., Goldman-Rakic, P.S., 1988. The alpha-2 adrenergic agonist guanfacine improves memory in aged monkeys without sedative or hypotensive side effects: evidence for alpha-2 receptor subtypes. J. Neurosci. Off. J. Soc. Neurosci. 8, 4287–4298.
Arnsten, A.F., Goldman-Rakic, P.S., 1985. Alpha 2-adrenergic mechanisms in prefrontal cortex associated with cognitive decline in aged nonhuman primates. Science 230, 1273–1276. https://doi.org/10.1126/science.2999977
Arnsten, A.F.T., 2006. Stimulants: Therapeutic actions in ADHD. Neuropsychopharmacol. Off. Publ. Am. Coll. Neuropsychopharmacol. 31, 2376–2383. https://doi.org/10.1038/sj.npp.1301164
Arnsten, A.F.T., 2000. Through the Looking Glass: Differential Noradenergic Modulation of Prefrontal Cortical Function. Neural Plast. 7, 133–146. https://doi.org/10.1155/NP.2000.133
Arnsten, A.F.T., Pliszka, S.R., 2011. Catecholamine influences on prefrontal cortical function: relevance to treatment of attention deficit/hyperactivity disorder and related disorders. Pharmacol. Biochem. Behav. 99, 211–216. https://doi.org/10.1016/j.pbb.2011.01.020
Arnsten, A.F.T., Rubia, K., 2012. Neurobiological Circuits Regulating Attention, Cognitive Control, Motivation, and Emotion: Disruptions in Neurodevelopmental Psychiatric Disorders. J. Am. Acad. Child Adolesc. Psychiatry 51, 356–367. https://doi.org/10.1016/j.jaac.2012.01.008
Arthur, D.L., Starr, A., 1984. Task-relevant late positive component of the auditory event-related potential in monkeys resembles P300 in humans. Science 223, 186–188. https://doi.org/10.1126/science.6691145
Ascending projections of the locus coeruleus in the rat. II. Autoradiographic study, 1977. . Brain Res. 127, 23–53. https://doi.org/10.1016/0006-8993(77)90378-X
Aston-Jones, G., 2005. Brain structures and receptors involved in alertness. Sleep Med. 6, S3–
S7. https://doi.org/10.1016/S1389-9457(05)80002-4
Aston-Jones, G., Bloom, F.E., 1981. Activity of norepinephrine-containing locus coeruleus neurons in behaving rats anticipates fluctuations in the sleep-waking cycle. J. Neurosci. Off. J. Soc. Neurosci. 1, 876–886.
Aston-Jones, G., Cohen, J.D., 2005. An integrative theory of locus coeruleus-norepinephrine function: adaptive gain and optimal performance. Annu. Rev. Neurosci. 28, 403–450. https://doi.org/10.1146/annurev.neuro.28.061604.135709
Aston-Jones, G., Ennis, M., Pieribone, V.A., Nickell, W.T., Shipley, M.T., 1986. The brain nucleus locus coeruleus: restricted afferent control of a broad efferent network. Science 234, 734–737. https://doi.org/10.1126/science.3775363
Aston-Jones, G., Foote, S.L., Segal, M., 1985. Impulse conduction properties of noradrenergic locus coeruleus axons projecting to monkey cerebrocortex. Neuroscience 15, 765–777. https://doi.org/10.1016/0306-4522(85)90077-6
Aston-Jones, G., Rajkowski, J., Cohen, J., 1999. Role of locus coeruleus in attention and behavioral flexibility. Biol. Psychiatry 46, 1309–1320. https://doi.org/10.1016/S0006-3223(99)00140-7
Aston-Jones, G., Rajkowski, J., Kubiak, P., Alexinsky, T., 1994. Locus coeruleus neurons in monkey are selectively activated by attended cues in a vigilance task. J. Neurosci. 14, 4467–4480. https://doi.org/10.1523/JNEUROSCI.14-07-04467.1994
Aston-Jones, G., Segal, M., Bloom, F.E., 1980. Brain aminergic axons exhibit marked variability in conduction velocity. Brain Res. 195, 215–222. https://doi.org/10.1016/0006-8993(80)90880-x
Aston-Jones, G., Shipley, M.T., Chouvet, G., Ennis, M., van Bockstaele, E., Pieribone, V., Shiekhattar, R., Akaoka, H., Drolet, G., Astier, B., Charléty, P., Valentino, R.J., Williams, J.T., 1991. Afferent regulation of locus coeruleus neurons: anatomy, physiology and pharmacology, in: Progress in Brain Research. Elsevier, pp. 47–75. https://doi.org/10.1016/S0079-6123(08)63799-1
References
266
Awasthi, V., Pratt, J., 1990. The Effects of Monetary Incentives on Effort and Decision Performance: The Role of Cognitive Characteristics. Account. Rev. 65, 797–811.
Backs, R.W., Ryan, A.M., 1992. Multimodal Measures of Mental Workload during Dual-Task Performance: Energetic Demands of Cognitive Processes. Proc. Hum. Factors Soc. Annu. Meet. 36, 1413–1417. https://doi.org/10.1177/154193129203601811
Backs, R.W., Seljos, K.A., 1994. Metabolic and cardiorespiratory measures of mental effort: the effects of level of difficulty in a working memory task. Int. J. Psychophysiol. 16, 57–68. https://doi.org/10.1016/0167-8760(94)90042-6
Balleine, B.W., Dickinson, A., 1998. Goal-directed instrumental action: contingency and incentive learning and their cortical substrates. Neuropharmacology 37, 407–419. https://doi.org/10.1016/S0028-3908(98)00033-1
Bari, A., Eagle, D.M., Mar, A.C., Robinson, E.S.J., Robbins, T.W., 2009. Dissociable effects of noradrenaline, dopamine, and serotonin uptake blockade on stop task performance in rats. Psychopharmacology (Berl.) 205, 273–283. https://doi.org/10.1007/s00213-009-1537-0
Bari, A., Xu, S., Pignatelli, M., Takeuchi, D., Feng, J., Li, Y., Tonegawa, S., 2020. Differential attentional control mechanisms by two distinct noradrenergic coeruleo-frontal cortical pathways. Proc. Natl. Acad. Sci. U. S. A. 117, 29080–29089. https://doi.org/10.1073/pnas.2015635117
Barraclough, D.J., Conroy, M.L., Lee, D., 2004. Prefrontal cortex and decision making in a mixed-strategy game. Nat. Neurosci. 7, 404–410. https://doi.org/10.1038/nn1209
Baumeister, R.F., 1984. Choking under pressure: self-consciousness and paradoxical effects of incentives on skillful performance. J. Pers. Soc. Psychol. 46, 610–620. https://doi.org/10.1037//0022-3514.46.3.610
Baumeister, R.F., Bratslavsky, E., Muraven, M., Tice, D.M., 1998. Ego depletion: is the active self a limited resource? J. Pers. Soc. Psychol. 74, 1252–1265. https://doi.org/10.1037//0022-3514.74.5.1252
Baumeister, R.F., Heatherton, T.F., 1996. Self-Regulation Failure: An Overview. Psychol. Inq. 7, 1–15. https://doi.org/10.1207/s15327965pli0701_1
Beach, L.R., Mitchell, T.R., 1978. A Contingency Model for the Selection of Decision Strategies. Acad. Manage. Rev. 3, 439–449. https://doi.org/10.2307/257535
Beatty, J., 1982. Task-evoked pupillary responses, processing load, and the structure of processing resources. Psychol. Bull. 91, 276–292. https://doi.org/10.1037/0033-2909.91.2.276
Beck, S.M., Locke, H.S., Savine, A.C., Jimura, K., Braver, T.S., 2010. Primary and secondary rewards differentially modulate neural activity dynamics during working memory. PloS One 5, e9251. https://doi.org/10.1371/journal.pone.0009251
Beilock, S., 2010. Choke: what the secrets of the brain reveal about getting it right when you have to, 1st Free Press hardcover ed. ed. Free Press, New York.
Benton, D., Brett, V., Brain, P.F., 1987. Glucose improves attention and reaction to frustration in children. Biol. Psychol. 24, 95–100. https://doi.org/10.1016/0301-0511(87)90016-0
Benton, D., Owens, D.S., Parker, P.Y., 1994. Blood glucose influences memory and attention in young adults. Neuropsychologia 32, 595–607. https://doi.org/10.1016/0028-3932(94)90147-3
Benton, D., Parker, P.Y., Donohoe, R.T., 1996. The supply of glucose to the brain and cognitive functioning. J. Biosoc. Sci. 28, 463–479. https://doi.org/10.1017/S0021932000022537
Berridge, C.W., España, R.A., 2000. Synergistic sedative effects of noradrenergic alpha(1)- and beta-receptor blockade on forebrain electroencephalographic and behavioral indices. Neuroscience 99, 495–505. https://doi.org/10.1016/s0306-4522(00)00215-3
Berridge, C.W., Foote, S.L., 1991. Effects of locus coeruleus activation on electroencephalographic activity in neocortex and hippocampus. J. Neurosci. Off. J. Soc. Neurosci. 11, 3135–3145.
Berridge, C.W., Waterhouse, B.D., 2003. The locus coeruleus-noradrenergic system: modulation of behavioral state and state-
References
267
dependent cognitive processes. Brain Res. Brain Res. Rev. 42, 33–84. https://doi.org/10.1016/s0165-0173(03)00143-7
Bettman, J.R., Johnson, E.J., Payne, J.W., 1990. A componential analysis of cognitive effort in choice. Organ. Behav. Hum. Decis. Process. 45, 111–139. https://doi.org/10.1016/0749-5978(90)90007-V
Bettman, J.R., Zins, M.A., 1979. Information Format and Choice Task Effects in Decision Making. J. Consum. Res. 6, 141–53.
Bing, G., Filer, D., Miller, J.C., Stone, E.A., 1991. Noradrenergic activation of immediate early genes in rat cerebral cortex. Mol. Brain Res. 11, 43–46. https://doi.org/10.1016/0169-328X(91)90019-T
Bing, G., Stone, E.A., Zhang, Y., Filer, D., 1992. Immunohistochemical studies of noradrenergic-induced expression of c-fos in the rat CNS. Brain Res. 592, 57–62. https://doi.org/10.1016/0006-8993(92)91658-2
Bingham, B., McFadden, K., Zhang, X., Bhatnagar, S., Beck, S., Valentino, R., 2011. Early adolescence as a critical window during which social stress distinctly alters behavior and brain norepinephrine activity. Neuropsychopharmacol. Off. Publ. Am. Coll. Neuropsychopharmacol. 36, 896–909. https://doi.org/10.1038/npp.2010.229
Birnbaum, S.G., Yuan, P.X., Wang, M., Vijayraghavan, S., Bloom, A.K., Davis, D.J., Gobeske, K.T., Sweatt, J.D., Manji, H.K., Arnsten, A.F.T., 2004. Protein kinase C overactivity impairs prefrontal cortical regulation of working memory. Science 306, 882–884. https://doi.org/10.1126/science.1100021
Boehler, C.N., Schevernels, H., Hopf, J.-M., Stoppel, C.M., Krebs, R.M., 2014. Reward prospect rapidly speeds up response inhibition via reactive control. Cogn. Affect. Behav. Neurosci. 14, 593–609. https://doi.org/10.3758/s13415-014-0251-5
Bogacz, R., Brown, E., Moehlis, J., Holmes, P., Cohen, J.D., 2006. The physics of optimal decision making: A formal analysis of models of performance in two-alternative forced-choice tasks. Psychol. Rev. 113, 700–765. https://doi.org/10.1037/0033-295X.113.4.700
Boksem, M.A.S., Meijman, T.F., Lorist, M.M., 2006. Mental fatigue, motivation and action monitoring. Biol. Psychol. 72, 123–132. https://doi.org/10.1016/j.biopsycho.2005.08.007
Boksem, M.A.S., Tops, M., 2008. Mental fatigue: Costs and benefits. Brain Res. Rev. 59, 125–139. https://doi.org/10.1016/j.brainresrev.2008.07.001
Bonnelle, V., Manohar, S., Behrens, T., Husain, M., 2016. Individual Differences in Premotor Brain Systems Underlie Behavioral Apathy. Cereb. Cortex 26, 807–819. https://doi.org/10.1093/cercor/bhv247
Bonner, S., Hastie, R., Sprinkle, G., Young, S., 2000. A Review of the Effects of Financial Incentives on Performance in Laboratory Tasks: Implications for Management Accounting. J. Manag. Account. Res. - J Manag Acc. Res 12, 19–64. https://doi.org/10.2308/jmar.2000.12.1.19
Borchert, R.J., Rittman, T., Passamonti, L., Ye, Z., Sami, S., Jones, S.P., Nombela, C., Vázquez Rodríguez, P., Vatansever, D., Rae, C.L., Hughes, L.E., Robbins, T.W., Rowe, J.B., 2016. Atomoxetine Enhances Connectivity of Prefrontal Networks in Parkinson’s Disease. Neuropsychopharmacology 41, 2171–2177. https://doi.org/10.1038/npp.2016.18
Borderies, N., Bornert, P., Gilardeau, S., Bouret, S., 2020. Pharmacological evidence for the implication of noradrenaline in effort. PLOS Biol. 18, e3000793. https://doi.org/10.1371/journal.pbio.3000793
Borg, G.A., 1982. Psychophysical bases of perceived exertion. Med. Sci. Sports Exerc. 14, 377–381.
Botvinick, M.M., 2007. Conflict monitoring and decision making: Reconciling two perspectives on anterior cingulate function. Cogn. Affect. Behav. Neurosci. 7, 356–366. https://doi.org/10.3758/CABN.7.4.356
Botvinick, M.M., Braver, T.S., Barch, D.M., Carter, C.S., Cohen, J.D., 2001. Conflict monitoring and cognitive control. Psychol. Rev. 108, 624–652. https://doi.org/10.1037/0033-295X.108.3.624
Botvinick, M.M., Cohen, J.D., 2014. The computational and neural basis of cognitive control: charted territory and new frontiers.
References
268
Cogn. Sci. 38, 1249–1285. https://doi.org/10.1111/cogs.12126
Botvinick, M.M., Cohen, J.D., Carter, C.S., 2004. Conflict monitoring and anterior cingulate cortex: an update. Trends Cogn. Sci. 8, 539–546. https://doi.org/10.1016/j.tics.2004.10.003
Boureau, Y.-L., Sokol-Hessner, P., Daw, N.D., 2015. Deciding How To Decide: Self-Control and Meta-Decision Making. Trends Cogn. Sci. 19, 700–710. https://doi.org/10.1016/j.tics.2015.08.013
Bouret, S., Duvel, A., Onat, S., Sara, S.J., 2003. Phasic activation of locus ceruleus neurons by the central nucleus of the amygdala. J. Neurosci. Off. J. Soc. Neurosci. 23, 3491–3497.
Bouret, S., Ravel, S., Richmond, B.J., 2012. Complementary neural correlates of motivation in dopaminergic and noradrenergic neurons of monkeys. Front. Behav. Neurosci. 6. https://doi.org/10.3389/fnbeh.2012.00040
Bouret, S., Richmond, B.J., 2015. Sensitivity of Locus Ceruleus Neurons to Reward Value for Goal-Directed Actions. J. Neurosci. 35, 4005–4014. https://doi.org/10.1523/JNEUROSCI.4553-14.2015
Bouret, S., Richmond, B.J., 2009. Relation of Locus Coeruleus Neurons in Monkeys to Pavlovian and Operant Behaviors. J. Neurophysiol. 101, 898–911. https://doi.org/10.1152/jn.91048.2008
Bouret, S., Sara, S.J., 2005. Network reset: a simplified overarching theory of locus coeruleus noradrenaline function. Trends Neurosci. 28, 574–582. https://doi.org/10.1016/j.tins.2005.09.002
Bouret, S., Sara, S.J., 2004. Reward expectation, orientation of attention and locus coeruleus-medial frontal cortex interplay during learning. Eur. J. Neurosci. 20, 791–802. https://doi.org/10.1111/j.1460-9568.2004.03526.x
Bouret, S., Sara, S.J., 2002. Locus coeruleus activation modulates firing rate and temporal organization of odour-induced single-cell responses in rat piriform cortex. Eur. J. Neurosci. 16, 2371–2382. https://doi.org/10.1046/j.1460-9568.2002.02413.x
Bowman, E.M., Aigner, T.G., Richmond, B.J., 1996. Neural signals in the monkey ventral striatum related to motivation for juice and cocaine rewards. J. Neurophysiol. 75, 1061–1073. https://doi.org/10.1152/jn.1996.75.3.1061
Braith R W, Wood C E, Limacher M C, Pollock M L, Lowenthal D T, Phillips M I, Staples E D, 1992. Abnormal neuroendocrine responses during exercise in heart transplant recipients. Circulation 86, 1453–1463. https://doi.org/10.1161/01.CIR.86.5.1453
Brink, R.L. van den, Pfeffer, T., Warren, C.M., Murphy, P.R., Tona, K.-D., Wee, N.J.A. van der, Giltay, E., Noorden, M.S. van, Rombouts, S.A.R.B., Donner, T.H., Nieuwenhuis, S., 2016. Catecholaminergic Neuromodulation Shapes Intrinsic MRI Functional Connectivity in the Human Brain. J. Neurosci. 36, 7865–7876. https://doi.org/10.1523/JNEUROSCI.0744-16.2016
Buffalari, D.M., Grace, A.A., 2007. Noradrenergic modulation of basolateral amygdala neuronal activity: opposing influences of alpha-2 and beta receptor activation. J. Neurosci. Off. J. Soc. Neurosci. 27, 12358–12366. https://doi.org/10.1523/JNEUROSCI.2007-07.2007
Burke, C.J., Brünger, C., Kahnt, T., Park, S.Q., Tobler, P.N., 2013. Neural Integration of Risk and Effort Costs by the Frontal Pole: Only upon Request. J. Neurosci. 33, 1706–1713. https://doi.org/10.1523/JNEUROSCI.3662-12.2013
Bush, G., Vogt, B.A., Holmes, J., Dale, A.M., Greve, D., Jenike, M.A., Rosen, B.R., 2002. Dorsal anterior cingulate cortex: A role in reward-based decision making. Proc. Natl. Acad. Sci. 99, 523–528. https://doi.org/10.1073/pnas.012470999
Bymaster, F.P., Katner, J.S., Nelson, D.L., Hemrick-Luecke, S.K., Threlkeld, P.G., Heiligenstein, J.H., Morin, S.M., Gehlert, D.R., Perry, K.W., 2002. Atomoxetine increases extracellular levels of norepinephrine and dopamine in prefrontal cortex of rat: a potential mechanism for efficacy in attention deficit/hyperactivity disorder. Neuropsychopharmacol. Off. Publ. Am. Coll. Neuropsychopharmacol. 27, 699–711. https://doi.org/10.1016/S0893-133X(02)00346-9
References
269
Cadegiani, F.A., Kater, C.E., 2019. Novel insights of overtraining syndrome discovered from the EROS study. BMJ Open Sport — Exerc. Med. 5. https://doi.org/10.1136/bmjsem-2019-000542
Cahill, L., McGaugh, J.L., 1996. Modulation of memory storage. Curr. Opin. Neurobiol. 6, 237–242. https://doi.org/10.1016/s0959-4388(96)80078-x
Camerer, C.F., Hogarth, R.M., 1999. The Effects of Financial Incentives in Experiments: A Review and Capital-Labor-Production Framework. J. Risk Uncertain. 19, 7–42. https://doi.org/10.1023/A:1007850605129
Carey, M.R., Regehr, W.G., 2009. Noradrenergic Control of Associative Synaptic Plasticity by Selective Modulation of Instructive Signals. Neuron 62, 112–122. https://doi.org/10.1016/j.neuron.2009.02.022
Carpenter, P.A., Just, M.A., 1975. Sentence comprehension: A psycholinguistic processing model of verification. Psychol. Rev. 82, 45–73. https://doi.org/10.1037/h0076248
Carroll, D., Turner, J.R., Hellawell, J.C., 1986a. Heart Rate and Oxygen Consumption during Active Psychological Challenge: The Effects of Level of Difficulty. Psychophysiology 23, 174–181. https://doi.org/10.1111/j.1469-8986.1986.tb00613.x
Carroll, D., Turner, J.R., Prasad, R., 1986b. The effects of level of difficulty of mental arithmetic challenge on heart rate and oxygen consumption. Int. J. Psychophysiol. 4, 167–173. https://doi.org/10.1016/0167-8760(86)90012-7
Carter, M.E., Yizhar, O., Chikahisa, S., Nguyen, H., Adamantidis, A., Nishino, S., Deisseroth, K., de Lecea, L., 2010. Tuning arousal with optogenetic modulation of locus coeruleus neurons. Nat. Neurosci. 13, 1526–1533. https://doi.org/10.1038/nn.2682
Cassens, G., Kuruc, A., Roffman, M., Orsulak, P.J., Schildkraut, J.J., 1981. Alterations in brain norepinephrine metabolism and behavior induced by environmental stimuli previously paired with inescapable shock. Behav. Brain Res. 2, 387–407. https://doi.org/10.1016/0166-4328(81)90020-6
Chamba, G., Weissmann, D., Rousset, C., Renaud, B., Pujol, J.F., 1991. Distribution of alpha-1 and
alpha-2 binding sites in the rat locus coeruleus. Brain Res. Bull. 26, 185–193. https://doi.org/10.1016/0361-9230(91)90225-9
Chamberlain, S.R., Del Campo, N., Dowson, J., Müller, U., Clark, L., Robbins, T.W., Sahakian, B.J., 2007. Atomoxetine improved response inhibition in adults with attention deficit/hyperactivity disorder. Biol. Psychiatry 62, 977–984. https://doi.org/10.1016/j.biopsych.2007.03.003
Chamberlain, S.R., Hampshire, A., Müller, U., Rubia, K., Del Campo, N., Craig, K., Regenthal, R., Suckling, J., Roiser, J.P., Grant, J.E., Bullmore, E.T., Robbins, T.W., Sahakian, B.J., 2009. Atomoxetine modulates right inferior frontal activation during inhibitory control: a pharmacological functional magnetic resonance imaging study. Biol. Psychiatry 65, 550–555. https://doi.org/10.1016/j.biopsych.2008.10.014
Chamberlain, S.R., Müller, U., Blackwell, A.D., Clark, L., Robbins, T.W., Sahakian, B.J., 2006a. Neurochemical modulation of response inhibition and probabilistic learning in humans. Science 311, 861–863. https://doi.org/10.1126/science.1121218
Chamberlain, S.R., Müller, U., Blackwell, A.D., Robbins, T.W., Sahakian, B.J., 2006b. Noradrenergic modulation of working memory and emotional memory in humans. Psychopharmacology (Berl.) 188, 397–407. https://doi.org/10.1007/s00213-006-0391-6
Chandler, D.J., Gao, W.-J., Waterhouse, B.D., 2014. Heterogeneous organization of the locus coeruleus projections to prefrontal and motor cortices. Proc. Natl. Acad. Sci. 111, 6816–6821. https://doi.org/10.1073/pnas.1320827111
Chaulk, P.C., Harley, C.W., 1998. Intracerebroventricular norepinephrine potentiation of the perforant path-evoked potential in dentate gyrus of anesthetized and awake rats: A role for both alpha- and beta-adrenoceptor activation. Brain Res. 787, 59–70. https://doi.org/10.1016/s0006-8993(97)01460-1
Chipman, P.H., Toma, J.S., Rafuse, V.F., 2012. Chapter 16 - Generation of motor neurons from pluripotent stem cells, in: Dunnett, S.B., Björklund, A. (Eds.), Progress in Brain Research, Functional Neural Transplantation III. Elsevier, pp. 313–331. https://doi.org/10.1016/B978-0-444-59544-7.00015-9
References
270
Chittka, L., Dyer, A.G., Bock, F., Dornhaus, A., 2003. Bees trade off foraging speed for accuracy. Nature 424, 388–388. https://doi.org/10.1038/424388a
Chittka, L., Skorupski, P., Raine, N.E., 2009. Speed–accuracy tradeoffs in animal decision making. Trends Ecol. Evol. 24, 400–407. https://doi.org/10.1016/j.tree.2009.02.010
Chong, T.T.-J., 2015. Disrupting the Perception of Effort with Continuous Theta Burst Stimulation. J. Neurosci. 35, 13269–13271. https://doi.org/10.1523/JNEUROSCI.2624-15.2015
Chong, T.T.-J., Bonnelle, V., Manohar, S., Veromann, K.-R., Muhammed, K., Tofaris, G.K., Hu, M., Husain, M., 2015. Dopamine enhances willingness to exert effort for reward in Parkinson’s disease. Cortex 69, 40–46. https://doi.org/10.1016/j.cortex.2015.04.003
Christi, M.J., Jelinek, H.F., 1993. Dye-coupling among neurons of the rat locus coeruleus during postnatal development. Neuroscience 56, 129–137. https://doi.org/10.1016/0306-4522(93)90568-Z
Christie, M.J., 1997. Generators of synchronous activity of the locus coeruleus during development. Semin. Cell Dev. Biol. 8, 29–34. https://doi.org/10.1006/scdb.1996.0118
Christie, M.J., Williams, J.T., North, R.A., 1989. Electrical coupling synchronizes subthreshold activity in locus coeruleus neurons in vitro from neonatal rats. J. Neurosci. 9, 3584–3589. https://doi.org/10.1523/JNEUROSCI.09-10-03584.1989
Christie, S.T., Schrater, P., 2015. Cognitive cost as dynamic allocation of energetic resources. Front. Neurosci. 9, 289. https://doi.org/10.3389/fnins.2015.00289
Ciombor, K.J., Ennis, M., Shipley, M.T., 1999. Norepinephrine increases rat mitral cell excitatory responses to weak olfactory nerve input via alpha-1 receptors in vitro. Neuroscience 90, 595–606. https://doi.org/10.1016/S0306-4522(98)00437-0
Citri, A., Malenka, R.C., 2008. Synaptic Plasticity: Multiple Forms, Functions, and Mechanisms. Neuropsychopharmacology 33, 18–41. https://doi.org/10.1038/sj.npp.1301559
Clark, C.R., Geffen, G.M., Geffen, L.B., 1987. Catecholamines and attention. I: Animal and clinical studies. Neurosci. Biobehav. Rev. 11, 341–352. https://doi.org/10.1016/s0149-7634(87)80006-4
Clark, L., Cools, R., Robbins, T.W., 2004. The neuropsychology of ventral prefrontal cortex: decision-making and reversal learning. Brain Cogn. 55, 41–53. https://doi.org/10.1016/S0278-2626(03)00284-7
Clayson, P.E., Larson, M.J., 2013. Psychometric properties of conflict monitoring and conflict adaptation indices: Response time and conflict N2 event-related potentials. Psychophysiology 50, 1209–1219. https://doi.org/10.1111/psyp.12138
Clayton, E.C., Rajkowski, J., Cohen, J.D., Aston-Jones, G., 2004. Phasic Activation of Monkey Locus Ceruleus Neurons by Simple Decisions in a Forced-Choice Task. J. Neurosci. 24, 9914–9920. https://doi.org/10.1523/JNEUROSCI.2446-04.2004
Clemow, D.B., Bushe, C.J., 2015. Atomoxetine in patients with ADHD: A clinical and pharmacological review of the onset, trajectory, duration of response and implications for patients. J. Psychopharmacol. Oxf. Engl. 29, 1221–1230. https://doi.org/10.1177/0269881115602489
Cléry-Melin, M.-L., Schmidt, L., Lafargue, G., Baup, N., Fossati, P., Pessiglione, M., 2011. Why Don’t You Try Harder? An Investigation of Effort Production in Major Depression. PLOS ONE 6, e23178. https://doi.org/10.1371/journal.pone.0023178
Cleveland, C.J., Ayres, R.U. (Eds.), 2004. Encyclopedia of energy. Elsevier Academic Press, Amsterdam ; Boston.
Coggan, J.S., Keller, D., Calì, C., Lehväslaiho, H., Markram, H., Schürmann, F., Magistretti, P.J., 2018. Norepinephrine stimulates glycogenolysis in astrocytes to fuel neurons with lactate. PLoS Comput. Biol. 14, e1006392. https://doi.org/10.1371/journal.pcbi.1006392
Cohen, J., Dunbar, K., Mcclelland, J., 1990. Cohen JD, McClelland JL, Dunbar K. On the control of automatic processes: a parallel distributed processing account of the Stroop effect. Psychol
References
271
Rev 97: 332-361. Psychol. Rev. 97, 332–61. https://doi.org/10.1037/0033-295X.97.3.332
Cohen, R., Lohr, I., Paul, R., Boland, R., 2001. Impairments of attention and effort among patients with major affective disorders. J. Neuropsychiatry Clin. Neurosci. 13, 385–395. https://doi.org/10.1176/jnp.13.3.385
Cohen, Z., Molinatti, G., Hamel, E., 1997. Astroglial and vascular interactions of noradrenaline terminals in the rat cerebral cortex. J. Cereb. Blood Flow Metab. Off. J. Int. Soc. Cereb. Blood Flow Metab. 17, 894–904. https://doi.org/10.1097/00004647-199708000-00008
Coradazzi, M., Gulino, R., Fieramosca, F., Falzacappa, L.V., Riggi, M., Leanza, G., 2016. Selective noradrenaline depletion impairs working memory and hippocampal neurogenesis. Neurobiol. Aging 48, 93–102. https://doi.org/10.1016/j.neurobiolaging.2016.08.012
Corbetta, M., Patel, G., Shulman, G.L., 2008. The Reorienting System of the Human Brain: From Environment to Theory of Mind. Neuron 58, 306–324. https://doi.org/10.1016/j.neuron.2008.04.017
Coricelli, G., Critchley, H.D., Joffily, M., O’Doherty, J.P., Sirigu, A., Dolan, R.J., 2005. Regret and its avoidance: a neuroimaging study of choice behavior. Nat. Neurosci. 8, 1255–1262. https://doi.org/10.1038/nn1514
Coull, J.T., Büchel, C., Friston, K.J., Frith, C.D., 1999. Noradrenergically mediated plasticity in a human attentional neuronal network. NeuroImage 10, 705–715. https://doi.org/10.1006/nimg.1999.0513
Coull, J.T., Nobre, A.C., Frith, C.D., 2001. The Noradrenergic α2 Agonist Clonidine Modulates Behavioural and Neuroanatomical Correlates of Human Attentional Orienting and Alerting. Cereb. Cortex 11, 73–84. https://doi.org/10.1093/cercor/11.1.73
Cousins, M.S., Salamone, J.D., 1994. Nucleus accumbens dopamine depletions in rats affect relative response allocation in a novel cost/benefit procedure. Pharmacol. Biochem. Behav. 49, 85–91. https://doi.org/10.1016/0091-3057(94)90460-x
Cowan, N., 2005. Working memory capacity, Essays in cognitive psychology. Psychology Press, New York.
Croxson, P.L., Walton, M.E., O’Reilly, J.X., Behrens, T.E.J., Rushworth, M.F.S., 2009. Effort-Based Cost–Benefit Valuation and the Human Brain. J. Neurosci. 29, 4531–4541. https://doi.org/10.1523/JNEUROSCI.4515-08.2009
Culbreth, A., Westbrook, A., Barch, D., 2016. Negative symptoms are associated with an increased subjective cost of cognitive effort. J. Abnorm. Psychol. 125, 528–536. https://doi.org/10.1037/abn0000153
Curtis, A.L., Lechner, S.M., Pavcovich, L.A., Valentino, R.J., 1997. Activation of the locus coeruleus noradrenergic system by intracoerulear microinfusion of corticotropin-releasing factor: effects on discharge rate, cortical norepinephrine levels and cortical electroencephalographic activity. J. Pharmacol. Exp. Ther. 281, 163–172.
Curtis, A.L., Leiser, S.C., Snyder, K., Valentino, R.J., 2012. Predator stress engages corticotropin-releasing factor and opioid systems to alter the operating mode of locus coeruleus norepinephrine neurons. Neuropharmacology 62, 1737–1745. https://doi.org/10.1016/j.neuropharm.2011.11.020
Da Silva, D.F., Mohammad, S., Hutchinson, K.A., Adamo, K.B., 2020. Cross-Validation of Ratings of Perceived Exertion Derived from Heart Rate Target Ranges Recommended for Pregnant Women. Int. J. Exerc. Sci. 13, 1340–1351.
Dahlstroem, A., Fuxe, K., 1964. EVIDENCE FOR THE EXISTENCE OF MONOAMINE-CONTAINING NEURONS IN THE CENTRAL NERVOUS SYSTEM. I. DEMONSTRATION OF MONOAMINES IN THE CELL BODIES OF BRAIN STEM NEURONS. Acta Physiol. Scand. Suppl. SUPPL 232:1-55.
Dajani, D.R., Uddin, L.Q., 2015. Demystifying cognitive flexibility: Implications for clinical and developmental neuroscience. Trends Neurosci. 38, 571–578. https://doi.org/10.1016/j.tins.2015.07.003
Dalton, G.L., Ma, L.M., Phillips, A.G., Floresco, S.B., 2011. Blockade of NMDA GluN2B receptors selectively impairs behavioral flexibility but not
References
272
initial discrimination learning. Psychopharmacology (Berl.) 216, 525–535. https://doi.org/10.1007/s00213-011-2246-z
Dan, Y., Poo, M.-M., 2006. Spike timing-dependent plasticity: from synapse to perception. Physiol. Rev. 86, 1033–1048. https://doi.org/10.1152/physrev.00030.2005
Dayan, P., 2012. How to set the switches on this thing. Curr. Opin. Neurobiol., Decision making 22, 1068–1074. https://doi.org/10.1016/j.conb.2012.05.011
de Gee, J.W., Colizoli, O., Kloosterman, N.A., Knapen, T., Nieuwenhuis, S., Donner, T.H., 2017. Dynamic modulation of decision biases by brainstem arousal systems. eLife 6, e23232. https://doi.org/10.7554/eLife.23232
de Gee, J.W., Knapen, T., Donner, T.H., 2014. Decision-related pupil dilation reflects upcoming choice and individual bias. Proc. Natl. Acad. Sci. 111, E618–E625. https://doi.org/10.1073/pnas.1317557111
de Gee, J.W., Tsetsos, K., Schwabe, L., Urai, A.E., McCormick, D., McGinley, M.J., Donner, T.H., 2020. Pupil-linked phasic arousal predicts a reduction of choice bias across species and decision domains. eLife 9, e54014. https://doi.org/10.7554/eLife.54014
De Martino, B., Kumaran, D., Seymour, B., Dolan, R.J., 2006. Frames, biases, and rational decision-making in the human brain. Science 313, 684–687. https://doi.org/10.1126/science.1128356
de Morree, H.M., Klein, C., Marcora, S.M., 2012. Perception of effort reflects central motor command during movement execution. Psychophysiology 49, 1242–1253. https://doi.org/10.1111/j.1469-8986.2012.01399.x
De Neys, W., Glumicic, T., 2008. Conflict monitoring in dual process theories of thinking. Cognition 106, 1248–1299. https://doi.org/10.1016/j.cognition.2007.06.002
De Neys, W., Vartanian, O., Goel, V., 2008. Smarter than we think: when our brains detect that we are biased. Psychol. Sci. 19, 483–489. https://doi.org/10.1111/j.1467-9280.2008.02113.x
de Rover, M., Brown, S.B.R.E., Band, G.P., Giltay, E.J., van Noorden, M.S., van der Wee, N.J.A., Nieuwenhuis, S., 2015. Beta receptor-mediated modulation of the oddball P3 but not error-related ERP components in humans. Psychopharmacology (Berl.) 232, 3161–3172. https://doi.org/10.1007/s00213-015-3966-2
De Sarro, G.B., Ascioti, C., Froio, F., Libri, V., Nisticò, G., 1987. Evidence that locus coeruleus is the site where clonidine and drugs acting at alpha 1- and alpha 2-adrenoceptors affect sleep and arousal mechanisms. Br. J. Pharmacol. 90, 675–685. https://doi.org/10.1111/j.1476-5381.1987.tb11220.x
Dehaene, S., Posner, M.I., Tucker, D.M., 1994. Localization of a Neural System for Error Detection and Compensation. Psychol. Sci. 5, 303–305. https://doi.org/10.1111/j.1467-9280.1994.tb00630.x
Demonstration of extensive brainstem projections to medial and lateral thalamus and hypothalamus in the rat, 1990. . Neuroscience 35, 609–626. https://doi.org/10.1016/0306-4522(90)90333-Y
Denes-Raj, V., Epstein, S., 1994. Conflict between intuitive and rational processing: When people behave against their better judgment. J. Pers. Soc. Psychol. 66, 819–829. https://doi.org/10.1037/0022-3514.66.5.819
Denson, T.F., von Hippel, W., Kemp, R.I., Teo, L.S., 2010. Glucose consumption decreases impulsive aggression in response to provocation in aggressive individuals. J. Exp. Soc. Psychol. 46, 1023–1028. https://doi.org/10.1016/j.jesp.2010.05.023
Devauges, V., Sara, S.J., 1990. Activation of the noradrenergic system facilitates an attentional shift in the rat. Behav. Brain Res. 39, 19–28. https://doi.org/10.1016/0166-4328(90)90118-x
Devilbiss, D.M., Page, M.E., Waterhouse, B.D., 2006. Locus ceruleus regulates sensory encoding by neurons and networks in waking animals. J. Neurosci. Off. J. Soc. Neurosci. 26, 9860–9872. https://doi.org/10.1523/JNEUROSCI.1776-06.2006
Devoto, P., Flore, G., 2006. On the origin of cortical dopamine: is it a co-transmitter in noradrenergic neurons? Curr. Neuropharmacol.
References
273
4, 115–125. https://doi.org/10.2174/157015906776359559
Devoto, P., Flore, G., Longu, G., Pira, L., Gessa, G.L., 2003. Origin of extracellular dopamine from dopamine and noradrenaline neurons in the medial prefrontal and occipital cortex. Synap. N. Y. N 50, 200–205. https://doi.org/10.1002/syn.10264
Devoto, P., Flore, G., Pani, L., Gessa, G.L., 2001. Evidence for co-release of noradrenaline and dopamine from noradrenergic neurons in the cerebral cortex. Mol. Psychiatry 6, 657–664. https://doi.org/10.1038/sj.mp.4000904
Devoto, P., Flore, G., Saba, P., Fà, M., Gessa, G.L., 2005. Co-release of noradrenaline and dopamine in the cerebral cortex elicited by single train and repeated train stimulation of the locus coeruleus. BMC Neurosci. 6, 31. https://doi.org/10.1186/1471-2202-6-31
Dewey, J., 1897. The Psychology of Effort. Philos. Rev. 6, 43–56. https://doi.org/10.2307/2175586
Dickinson, A., 1980. Contemporary Animal Learning Theory. CUP Archive.
Dickinson, A., Balleine, B., 1994. Motivational control of goal-directed action. Anim. Learn. Behav. 22, 1–18. https://doi.org/10.3758/BF03199951
Dienel, G.A., 2017. The metabolic trinity, glucose-glycogen-lactate, links astrocytes and neurons in brain energetics, signaling, memory, and gene expression. Neurosci. Lett. 637, 18–25. https://doi.org/10.1016/j.neulet.2015.02.052
Dixon, M.L., Christoff, K., 2012. The Decision to Engage Cognitive Control Is Driven by Expected Reward-Value: Neural and Behavioral Evidence. PLOS ONE 7, e51637. https://doi.org/10.1371/journal.pone.0051637
Dmochowski, J.P., Norcia, A.M., 2015. Cortical Components of Reaction-Time during Perceptual Decisions in Humans. PLOS ONE 10, e0143339. https://doi.org/10.1371/journal.pone.0143339
Donohoe, R.T., Benton, D., 1999. Blood glucose control and aggressiveness in females. Personal. Individ. Differ. 26, 905–911. https://doi.org/10.1016/S0191-8869(98)00191-3
Dorr, A.E., Debonnel, G., 2006. Effect of vagus nerve stimulation on serotonergic and noradrenergic transmission. J. Pharmacol. Exp. Ther. 318, 890–898. https://doi.org/10.1124/jpet.106.104166
Dow, J., Werlang, S.R. da C., 1992. Uncertainty Aversion, Risk Aversion, and the Optimal Choice of Portfolio. Econometrica 60, 197–204. https://doi.org/10.2307/2951685
Doya, K., 2002. Metalearning and neuromodulation. Neural Netw. 15, 495–506. https://doi.org/10.1016/S0893-6080(02)00044-8
Dretsch, M.N., Tipples, J., 2008. Working memory involved in predicting future outcomes based on past experiences. Brain Cogn. 66, 83–90. https://doi.org/10.1016/j.bandc.2007.05.006
Drugowitsch, J., Wyart, V., Devauchelle, A.-D., Koechlin, E., 2016. Computational Precision of Mental Inference as Critical Source of Human Choice Suboptimality. Neuron 92, 1398–1411. https://doi.org/10.1016/j.neuron.2016.11.005
Dubois, M., Habicht, J., Michely, J., Moran, R., Dolan, R.J., Hauser, T.U., 2021. Human complex exploration strategies are enriched by noradrenaline-modulated heuristics. eLife 10, e59907. https://doi.org/10.7554/eLife.59907
Dunn, T.L., Lutes, D.J.C., Risko, E.F., 2016. Metacognitive evaluation in the avoidance of demand. J. Exp. Psychol. Hum. Percept. Perform. 42, 1372–1387. https://doi.org/10.1037/xhp0000236
Dvorak, R.D., Simons, J.S., 2009. Moderation of Resource Depletion in the Self-Control Strength Model: Differing Effects of Two Modes of Self-Control. Pers. Soc. Psychol. Bull. 35, 572–583. https://doi.org/10.1177/0146167208330855
Einhäuser, W., Koch, C., Carter, O.L., 2010. Pupil Dilation Betrays the Timing of Decisions. Front. Hum. Neurosci. 4, 18. https://doi.org/10.3389/fnhum.2010.00018
Einhauser, W., Stout, J., Koch, C., Carter, O., 2008. Pupil dilation reflects perceptual selection and predicts subsequent stability in perceptual rivalry. Proc. Natl. Acad. Sci. 105, 1704–1709. https://doi.org/10.1073/pnas.0707727105
References
274
Ellis, K.A., Nathan, P.J., 2001. The pharmacology of human working memory. Int. J. Neuropsychopharmacol. 4, 299–313. https://doi.org/10.1017/S1461145701002541
Engelmann, J.B., Damaraju, E., Padmala, S., Pessoa, L., 2009. Combined Effects of Attention and Motivation on Visual Task Performance: Transient and Sustained Motivational Effects. Front. Hum. Neurosci. 3. https://doi.org/10.3389/neuro.09.004.2009
Ennis, M., Aston-Jones, G., 1988. Activation of locus coeruleus from nucleus paragigantocellularis: a new excitatory amino acid pathway in brain. J. Neurosci. 8, 3644–3657. https://doi.org/10.1523/JNEUROSCI.08-10-03644.1988
Ennis, M., Aston-Jones, G., 1986. Evidence for self- and neighbor-mediated postactivation inhibition of locus coeruleus neurons. Brain Res. 374, 299–305. https://doi.org/10.1016/0006-8993(86)90424-5
Epstein, S., 1994. Integration of the cognitive and the psychodynamic unconscious. Am. Psychol. 49, 709–724. https://doi.org/10.1037/0003-066X.49.8.709
Epstein, S., Pacini, R., 1999. Some basic issues regarding dual-process theories from the perspective of cognitive–experiential self-theory, in: Dual-Process Theories in Social Psychology. The Guilford Press, New York, NY, US, pp. 462–482.
Ericsson, K.A., Simon, H.A., 1980. Verbal reports as data. Psychol. Rev. 87, 215–251. https://doi.org/10.1037/0033-295X.87.3.215
Eslinger, P.J., Damasio, A.R., 1985. Severe disturbance of higher cognition after bilateral frontal lobe ablation: patient EVR. Neurology 35, 1731–1741. https://doi.org/10.1212/wnl.35.12.1731
Evans, J.S.B.T., 1984. Heuristic and analytic processes in reasoning*. Br. J. Psychol. 75, 451–468. https://doi.org/10.1111/j.2044-8295.1984.tb01915.x
Evans, J.S.B.T., Over, D.E., 2013. Rationality and Reasoning. Psychology Press.
Evans, M., Shaw, A., Thompson, E.A., Falk, S., Turton, P., Thompson, T., Sharp, D., 2007.
Decisions to use complementary and alternative medicine (CAM) by male cancer patients: information-seeking roles and types of evidence used. BMC Complement. Altern. Med. 7, 25. https://doi.org/10.1186/1472-6882-7-25
Fairclough, S.H., Houston, K., 2004. A metabolic measure of mental effort. Biol. Psychol. 66, 177–190. https://doi.org/10.1016/j.biopsycho.2003.10.001
Faisal, A.A., Selen, L.P.J., Wolpert, D.M., 2008. Noise in the nervous system. Nat. Rev. Neurosci. 9, 292–303. https://doi.org/10.1038/nrn2258
Falcone, R., Weintraub, D.B., Setogawa, T., Wittig, J.H., Chen, G., Richmond, B.J., 2019. Temporal Coding of Reward Value in Monkey Ventral Striatal Tonically Active Neurons. J. Neurosci. Off. J. Soc. Neurosci. 39, 7539–7550. https://doi.org/10.1523/JNEUROSCI.0869-19.2019
Falkenberg, L.E., Westerhausen, R., Specht, K., Hugdahl, K., 2012. Resting-state glutamate level in the anterior cingulate predicts blood-oxygen level-dependent response to cognitive control. Proc. Natl. Acad. Sci. 109, 5069–5073.
Falkenstein, M., Hoormann, J., Hohnsbein, J., 1999. ERP components in Go/Nogo tasks and their relation to inhibition. Acta Psychol. (Amst.) 101, 267–291. https://doi.org/10.1016/S0001-6918(99)00008-6
Faraone, S.V., Biederman, J., Spencer, T., Michelson, D., Adler, L., Reimherr, F., Seidman, L., 2005. Atomoxetine and Stroop Task Performance in Adult Attention-Deficit/Hyperactivity Disorder. J. Child Adolesc. Psychopharmacol. 15, 664–670. https://doi.org/10.1089/cap.2005.15.664
Feldman, J., Barshi, I., 2007. The Effects of Blood Glucose Levels on Cognitive Performance: A Review of the Literature [WWW Document]. Tech. Rep. NASA Moffett Field CA. URL https://ntrs.nasa.gov/citations/20070031714 (accessed 4.23.21).
Fernando, A.B.P., Economidou, D., Theobald, D.E., Zou, M.-F., Newman, A.H., Spoelder, M., Caprioli, D., Moreno, M., Hipólito, L., Aspinall, A.T., Robbins, T.W., Dalley, J.W., 2012. Modulation of high impulsivity and attentional performance in rats by selective direct and indirect dopaminergic and noradrenergic
References
275
receptor agonists. Psychopharmacology (Berl.) 219, 341–352. https://doi.org/10.1007/s00213-011-2408-z
Ferry, B., Roozendaal, B., McGaugh, J.L., 1999. Basolateral Amygdala Noradrenergic Influences on Memory Storage Are Mediated by an Interaction between β- and α1-Adrenoceptors. J. Neurosci. 19, 5119–5123. https://doi.org/10.1523/JNEUROSCI.19-12-05119.1999
Fiedler, S., Glöckner, A., 2012. The Dynamics of Decision Making in Risky Choice: An Eye-Tracking Analysis. Front. Psychol. 3. https://doi.org/10.3389/fpsyg.2012.00335
Fiehler, K., Ullsperger, M., Cramon, D.Y.V., 2004. Neural correlates of error detection and error correction: is there a common neuroanatomical substrate? Eur. J. Neurosci. 19, 3081–3087. https://doi.org/10.1111/j.0953-816X.2004.03414.x
Filevich, E., Kühn, S., Haggard, P., 2013. There Is No Free Won’t: Antecedent Brain Activity Predicts Decisions to Inhibit. PLOS ONE 8, e53053. https://doi.org/10.1371/journal.pone.0053053
Filevich, E., Kühn, S., Haggard, P., 2012. Intentional inhibition in human action: the power of “no.” Neurosci. Biobehav. Rev. 36, 1107–1118. https://doi.org/10.1016/j.neubiorev.2012.01.006
Fink, K., Velebit, J., Vardjan, N., Zorec, R., Kreft, M., 2021. Noradrenaline-induced l-lactate production requires d-glucose entry and transit through the glycogen shunt in single-cultured rat astrocytes. J. Neurosci. Res. 99, 1084–1098. https://doi.org/10.1002/jnr.24783
Fischer, R., Ventura-Bort, C., Hamm, A., Weymar, M., 2018. Transcutaneous vagus nerve stimulation (tVNS) enhances conflict-triggered adjustment of cognitive control. Cogn. Affect. Behav. Neurosci. 18, 680–693. https://doi.org/10.3758/s13415-018-0596-2
Flicker, C., Geyer, M.A., 1982. The hippocampus as a possible site of action for increased locomotion during intracerebral infusions of norepinephrine. Behav. Neural Biol. 34, 421–426. https://doi.org/10.1016/s0163-1047(82)91843-x
Floresco, S.B., Tse, M.T.L., Ghods-Sharifi, S., 2008. Dopaminergic and Glutamatergic Regulation of Effort- and Delay-Based Decision Making. Neuropsychopharmacology 33, 1966–1979. https://doi.org/10.1038/sj.npp.1301565
Folstein, J.R., Van Petten, C., 2008. Influence of cognitive control and mismatch on the N2 component of the ERP: a review. Psychophysiology 45, 152–170. https://doi.org/10.1111/j.1469-8986.2007.00602.x
Foote, S.L., Aston-Jones, G., Bloom, F.E., 1980. Impulse activity of locus coeruleus neurons in awake rats and monkeys is a function of sensory stimulation and arousal. Proc. Natl. Acad. Sci. U. S. A. 77, 3033–3037. https://doi.org/10.1073/pnas.77.5.3033
Foote, S.L., Bloom, F.E., Aston-Jones, G., 1983. Nucleus locus ceruleus: new evidence of anatomical and physiological specificity. Physiol. Rev. 63, 844–914. https://doi.org/10.1152/physrev.1983.63.3.844
Foote, S.L., Freedman, R., Oliver, A.P., 1975. Effects of putative neurotransmitters on neuronal activity in monkey auditory cortex. Brain Res. 86, 229–242. https://doi.org/10.1016/0006-8993(75)90699-x
Förster, J., Higgins, E.T., Bianco, A.T., 2003. Speed/accuracy decisions in task performance: Built-in trade-off or separate strategic concerns? Organ. Behav. Hum. Decis. Process. 90, 148–164. https://doi.org/10.1016/S0749-5978(02)00509-5
Forstmann, B.U., Brass, M., Koch, I., von Cramon, D.Y., 2006. Voluntary selection of task sets revealed by functional magnetic resonance imaging. J. Cogn. Neurosci. 18, 388–398. https://doi.org/10.1162/089892906775990589
Fox, M.D., Raichle, M.E., 2007. Spontaneous fluctuations in brain activity observed with functional magnetic resonance imaging. Nat. Rev. Neurosci. 8, 700–711. https://doi.org/10.1038/nrn2201
Frangos, E., Ellrich, J., Komisaruk, B.R., 2015. Non-invasive Access to the Vagus Nerve Central Projections via Electrical Stimulation of the External Ear: fMRI Evidence in Humans. Brain Stimulat. 8, 624–636. https://doi.org/10.1016/j.brs.2014.11.018
References
276
Frank, M.J., Badre, D., 2012. Mechanisms of hierarchical reinforcement learning in corticostriatal circuits 1: computational analysis. Cereb. Cortex N. Y. N 1991 22, 509–526. https://doi.org/10.1093/cercor/bhr114
Franks, N.R., Dornhaus, A., Fitzsimmons, J.P., Stevens, M., 2003. Speed versus accuracy in collective decision making. Proc. R. Soc. Lond. B Biol. Sci. 270, 2457–2463. https://doi.org/10.1098/rspb.2003.2527
Franowicz, J.S., Arnsten, A.F., 1998. The alpha-2a noradrenergic agonist, guanfacine, improves delayed response performance in young adult rhesus monkeys. Psychopharmacology (Berl.) 136, 8–14. https://doi.org/10.1007/s002130050533
Freedman, R., Hoffer, B.J., Woodward, D.J., Puro, D., 1977. Interaction of norepinephrine with cerebellar activity evoked by mossy and climbing fibers. Exp. Neurol. 55, 269–288. https://doi.org/10.1016/0014-4886(77)90175-3
Fries, P., 2005. A mechanism for cognitive dynamics: neuronal communication through neuronal coherence. Trends Cogn. Sci. 9, 474–480. https://doi.org/10.1016/j.tics.2005.08.011
Fudenberg, D., Strack, P., Strzalecki, T., 2018. Speed, Accuracy, and the Optimal Timing of Choices. Am. Econ. Rev. 108, 3651–3684. https://doi.org/10.1257/aer.20150742
Fuxe, K., Borroto-Escuela, D.O., Romero-Fernandez, W., Diaz-Cabiale, Z., Rivera, A., Ferraro, L., Tanganelli, S., Tarakanov, A.O., Garriga, P., Narváez, J.A., Ciruela, F., Guescini, M., Agnati, L.F., 2012. Extrasynaptic neurotransmission in the modulation of brain function. Focus on the striatal neuronal-glial networks. Front. Physiol. 3, 136. https://doi.org/10.3389/fphys.2012.00136
Gaillard, A.W.K., 2001. Stress, workload, and fatigue as three biobehavioral states: A general overview, in: Stress, Workload, and Fatigue, Human Factors in Transportation. Lawrence Erlbaum Associates Publishers, Mahwah, NJ, US, pp. 623–639.
Gaillard, A.W.K., 1993. Comparing the concepts of mental load and stress. Ergonomics 36, 991–1005. https://doi.org/10.1080/00140139308967972
Gailliot, M.T., 2008. Unlocking the Energy Dynamics of Executive Functioning: Linking Executive Functioning to Brain Glycogen. Perspect. Psychol. Sci. J. Assoc. Psychol. Sci. 3, 245–263. https://doi.org/10.1111/j.1745-6924.2008.00077.x
Gailliot, M.T., Baumeister, R.F., 2007. The Physiology of Willpower: Linking Blood Glucose to Self-Control. Personal. Soc. Psychol. Rev. 11, 303–327. https://doi.org/10.1177/1088868307303030
Gailliot, M.T., Baumeister, R.F., DeWall, C.N., Maner, J.K., Plant, E.A., Tice, D.M., Brewer, L.E., Schmeichel, B.J., 2007. Self-control relies on glucose as a limited energy source: willpower is more than a metaphor. J. Pers. Soc. Psychol. 92, 325–336. https://doi.org/10.1037/0022-3514.92.2.325
Gajewski, P.D., Falkenstein, M., 2011. Diversity of the P3 in the task-switching paradigm. Brain Res. 1411, 87–97. https://doi.org/10.1016/j.brainres.2011.07.010
Gallagher, K.M., Fadel, P.J., Strømstad, M., Ide, K., Smith, S.A., Querry, R.G., Raven, P.B., Secher, N.H., 2001. Effects of partial neuromuscular blockade on carotid baroreflex function during exercise in humans. J. Physiol. 533, 861–870. https://doi.org/10.1111/j.1469-7793.2001.t01-1-00861.x
Gamo, N.J., Wang, M., Arnsten, A.F.T., 2010. Methylphenidate and Atomoxetine Enhance Prefrontal Function Through α2-Adrenergic and Dopamine D1 Receptors. J. Am. Acad. Child Adolesc. Psychiatry 49, 1011–1023. https://doi.org/10.1016/j.jaac.2010.06.015
Garavan, H., Ross, T.J., Stein, E.A., 1999. Right hemispheric dominance of inhibitory control: An event-related functional MRI study. Proc. Natl. Acad. Sci. 96, 8301–8306. https://doi.org/10.1073/pnas.96.14.8301
Garrison, K.E., Finley, A.J., Schmeichel, B.J., 2019. Ego Depletion Reduces Attention Control: Evidence From Two High-Powered Preregistered Experiments. Pers. Soc. Psychol. Bull. 45, 728–739. https://doi.org/10.1177/0146167218796473
Gatti, P.J., Hill, K.J., Da Silva, A.M., Norman, W.P., Gillis, R.A., 1988. Central nervous system site of
References
277
action for the hypotensive effect of clonidine in the cat. J. Pharmacol. Exp. Ther. 245, 373–380.
Gehring, W.J., Knight, R.T., 2000. Prefrontal–cingulate interactions in action monitoring. Nat. Neurosci. 3, 516–520. https://doi.org/10.1038/74899
Gelbard-Sagiv, H., Magidov, E., Sharon, H., Hendler, T., Nir, Y., 2018. Noradrenaline Modulates Visual Perception and Late Visually Evoked Activity. Curr. Biol. 28, 2239-2249.e6. https://doi.org/10.1016/j.cub.2018.05.051
Genkova-Papazova, M., Petkova, B.P., Lazarova-Bakarova, M., Boyanova, E., Staneva-Stoytcheva, D., 1997. Effects of F lunarizine and Nitrendipine on Electroconvulsive Shock- and Clonidine-Induced Amnesia. Pharmacol. Biochem. Behav. 56, 583–587. https://doi.org/10.1016/S0091-3057(96)00406-6
Geurts, H.M., Verté, S., Oosterlaan, J., Roeyers, H., Sergeant, J.A., 2004. How specific are executive functioning deficits in attention deficit hyperactivity disorder and autism? J. Child Psychol. Psychiatry 45, 836–854. https://doi.org/10.1111/j.1469-7610.2004.00276.x
Gibbs, M.E., 2015. Role of Glycogenolysis in Memory and Learning: Regulation by Noradrenaline, Serotonin and ATP. Front. Integr. Neurosci. 9, 70. https://doi.org/10.3389/fnint.2015.00070
Gibbs, M.E., Hutchinson, D.S., Summers, R.J., 2010. Noradrenaline release in the locus coeruleus modulates memory formation and consolidation; roles for α- and β-adrenergic receptors. Neuroscience 170, 1209–1222. https://doi.org/10.1016/j.neuroscience.2010.07.052
Gigerenzer, G., 2010. Rationality for mortals: how people cope with uncertainty, Evolution and cognition. Oxford University Press, New York ; Oxford.
Gigerenzer, G., 2008. Why Heuristics Work. Perspect. Psychol. Sci. 3, 20–29. https://doi.org/10.1111/j.1745-6916.2008.00058.x
Gigerenzer, G., Goldstein, D.G., 1996. Reasoning the fast and frugal way: models of bounded
rationality. Psychol. Rev. 103, 650–669. https://doi.org/10.1037/0033-295x.103.4.650
Gilden, D.L., 1997. Fluctuations in the Time Required for Elementary Decisions. Psychol. Sci. 8, 296–301. https://doi.org/10.1111/j.1467-9280.1997.tb00441.x
Gilzenrat, M.S., Holmes, B.D., Rajkowski, J., Aston-Jones, G., Cohen, J.D., 2002. Simplified dynamics in a model of noradrenergic modulation of cognitive performance. Neural Netw. 15, 647–663. https://doi.org/10.1016/S0893-6080(02)00055-2
Gilzenrat, M.S., Nieuwenhuis, S., Jepma, M., Cohen, J.D., 2010. Pupil diameter tracks changes in control state predicted by the adaptive gain theory of locus coeruleus function. Cogn. Affect. Behav. Neurosci. 10, 252–269. https://doi.org/10.3758/CABN.10.2.252
Gloeckner, A., Witteman, C., 2010. Beyond dual-process models: A categorisation of processes underlying intuitive judgement and decision making. Think. Reason. 16, 1–25. https://doi.org/10.1080/13546780903395748
Gneezy, U., Rustichini, A., 2000. Pay Enough or Don’t Pay at All. Q. J. Econ. 115, 791–810.
Goel, V., 1995. Sketches of Thought. MIT Press.
Goldman, G., Coleman, P.D., 1981. Neuron numbers in locus coeruleus do not change with age in Fisher 344 rat. Neurobiol. Aging 2, 33–36. https://doi.org/10.1016/0197-4580(81)90056-7
Grant, S.J., Aston-Jones, G., Redmond, D.E., 1988. Responses of primate locus coeruleus neurons to simple and complex sensory stimuli. Brain Res. Bull. 21, 401–410. https://doi.org/10.1016/0361-9230(88)90152-9
Green, L., Myerson, J., Oliveira, L., Chang, S.E., 2013. Delay discounting of monetary rewards over a wide range of amounts. J. Exp. Anal. Behav. 100, 269–281. https://doi.org/10.1002/jeab.45
Griffiths, T.L., Lieder, F., Goodman, N.D., 2015. Rational use of cognitive resources: levels of analysis between the computational and the algorithmic. Top. Cogn. Sci. 7, 217–229. https://doi.org/10.1111/tops.12142
References
278
Gross, J.J., 2011. Handbook of Emotion Regulation, First Edition. Guilford Press.
Gruetter, R., 2003. Glycogen: The forgotten cerebral energy store. J. Neurosci. Res. 74, 179–183. https://doi.org/10.1002/jnr.10785
Gubits, R.M., Smith, T.M., Fairhurst, J.L., Yu, H., 1989. Adrenergic receptors mediate changes in c-fos mRNA levels in brain. Mol. Brain Res. 6, 39–45. https://doi.org/10.1016/0169-328X(89)90026-0
Guedj, C., Meunier, D., Meunier, M., Hadj-Bouziane, F., 2017a. Could LC-NE-Dependent Adjustment of Neural Gain Drive Functional Brain Network Reorganization? Neural Plast. 2017, 4328015. https://doi.org/10.1155/2017/4328015
Guedj, C., Monfardini, E., Reynaud, A.J., Farnè, A., Meunier, M., Hadj-Bouziane, F., 2017b. Boosting Norepinephrine Transmission Triggers Flexible Reconfiguration of Brain Networks at Rest. Cereb. Cortex N. Y. N 1991 27, 4691–4700. https://doi.org/10.1093/cercor/bhw262
Habay, J., Van Cutsem, J., Verschueren, J., De Bock, S., Proost, M., De Wachter, J., Tassignon, B., Meeusen, R., Roelands, B., 2021. Mental Fatigue and Sport-Specific Psychomotor Performance: A Systematic Review. Sports Med. https://doi.org/10.1007/s40279-021-01429-6
Hagena, H., Hansen, N., Manahan-Vaughan, D., 2016. β-Adrenergic Control of Hippocampal Function: Subserving the Choreography of Synaptic Information Storage and Memory. Cereb. Cortex N. Y. NY 26, 1349–1364. https://doi.org/10.1093/cercor/bhv330
Hagger, M.S., Wood, C., Stiff, C., Chatzisarantis, N.L.D., 2010. Ego depletion and the strength model of self-control: A meta-analysis. Psychol. Bull. 136, 495–525. https://doi.org/10.1037/a0019486
Halliday, R., Naylor, H., Brandeis, D., Callaway, E., Yano, L., Herzig, K., 1994. The effect of D-amphetamine, clonidine, and yohimbine on human information processing. Psychophysiology 31, 331–337. https://doi.org/10.1111/j.1469-8986.1994.tb02441.x
Hare, T.A., Schultz, W., Camerer, C.F., O’Doherty, J.P., Rangel, A., 2011. Transformation of stimulus
value signals into motor commands during simple choice. Proc. Natl. Acad. Sci. 108, 18120–18125. https://doi.org/10.1073/pnas.1109322108
Harley, C.W., Sara, S.J., 1992. Locus coeruleus bursts induced by glutamate trigger delayed perforant path spike amplitude potentiation in the dentate gyrus. Exp. Brain Res. 89, 581–587. https://doi.org/10.1007/BF00229883
Hay, N., Russell, S., Tolpin, D., Shimony, S.E., 2014. Selecting Computations: Theory and Applications. ArXiv14082048 Cs.
Hayden, B.Y., Platt, M.L., 2010. Neurons in Anterior Cingulate Cortex Multiplex Information about Reward and Action. J. Neurosci. 30, 3339–3346.
Hazy, T.E., Frank, M.J., O’Reilly, R.C., 2007. Towards an executive without a homunculus: computational models of the prefrontal cortex/basal ganglia system. Philos. Trans. R. Soc. B Biol. Sci. 362, 1601–1613. https://doi.org/10.1098/rstb.2007.2055
Heath, T.P., Melichar, J.K., Nutt, D.J., Donaldson, L.F., 2006. Human Taste Thresholds Are Modulated by Serotonin and Noradrenaline. J. Neurosci. 26, 12664–12671. https://doi.org/10.1523/JNEUROSCI.3459-06.2006
Heginbotham, L.R., Dunwiddie, T.V., 1991. Long-term increases in the evoked population spike in the CA1 region of rat hippocampus induced by beta-adrenergic receptor activation. J. Neurosci. 11, 2519–2527. https://doi.org/10.1523/JNEUROSCI.11-08-02519.1991
Heitz, R.P., 2014. The speed-accuracy tradeoff: history, physiology, methodology, and behavior. Front. Neurosci. 8. https://doi.org/10.3389/fnins.2014.00150
Heitz, R.P., Schall, J.D., 2012. Neural Mechanisms of Speed-Accuracy Tradeoff. Neuron 76, 616–628. https://doi.org/10.1016/j.neuron.2012.08.030
Heitz, R.P., Schrock, J.C., Payne, T.W., Engle, R.W., 2007. Effects of incentive on working memory capacity: Behavioral and pupillometric data. Psychophysiology 0, 071003012229005-??? https://doi.org/10.1111/j.1469-8986.2007.00605.x
References
279
Hermans, E.J., Henckens, M.J.A.G., Joëls, M., Fernández, G., 2014. Dynamic adaptation of large-scale brain networks in response to acute stressors. Trends Neurosci. 37, 304–314. https://doi.org/10.1016/j.tins.2014.03.006
Hermans, E.J., Marle, H.J.F. van, Ossewaarde, L., Henckens, M.J.A.G., Qin, S., Kesteren, M.T.R. van, Schoots, V.C., Cousijn, H., Rijpkema, M., Oostenveld, R., Fernández, G., 2011. Stress-Related Noradrenergic Activity Prompts Large-Scale Neural Network Reconfiguration. Science 334, 1151–1153. https://doi.org/10.1126/science.1209603
Hernaus, D., Casales Santa, M.M., Offermann, J.S., Van Amelsvoort, T., 2017. Noradrenaline transporter blockade increases fronto-parietal functional connectivity relevant for working memory. Eur. Neuropsychopharmacol. 27, 399–410. https://doi.org/10.1016/j.euroneuro.2017.02.004
Hertz, L., Lovatt, D., Goldman, S.A., Nedergaard, M., 2010. Adrenoceptors in brain: cellular gene expression and effects on astrocytic metabolism and [Ca(2+)]i. Neurochem. Int. 57, 411–420. https://doi.org/10.1016/j.neuint.2010.03.019
Hertz, L., Peng, L., Dienel, G.A., 2007. Energy metabolism in astrocytes: high rate of oxidative metabolism and spatiotemporal dependence on glycolysis/glycogenolysis. J. Cereb. Blood Flow Metab. Off. J. Int. Soc. Cereb. Blood Flow Metab. 27, 219–249. https://doi.org/10.1038/sj.jcbfm.9600343
Hervé-Minvielle, A., Sara, S., 1995. Rapid habituation of auditory responses of locus coeruleus cells in anaesthetized and awake rats. Neuroreport 6, 1363–1368. https://doi.org/10.1097/00001756-199507100-00001
Hess, E.H., Polt, J.M., 1964. Pupil Size in Relation to Mental Activity during Simple Problem-Solving. Science 143, 1190–1192. https://doi.org/10.1126/science.143.3611.1190
Hillman, K.L., Bilkey, D.K., 2010. Neurons in the Rat Anterior Cingulate Cortex Dynamically Encode Cost–Benefit in a Spatial Decision-Making Task. J. Neurosci. 30, 7705–7713. https://doi.org/10.1523/JNEUROSCI.1273-10.2010
Hinson, J.M., Jameson, T.L., Whitney, P., 2003. Impulsive decision making and working memory. J. Exp. Psychol. Learn. Mem. Cogn. 29, 298–306. https://doi.org/10.1037/0278-7393.29.2.298
Hirschberg, S., Li, Y., Randall, A., Kremer, E.J., Pickering, A.E., 2017. Functional dichotomy in spinal- vs prefrontal-projecting locus coeruleus modules splits descending noradrenergic analgesia from ascending aversion and anxiety in rats. eLife 6. https://doi.org/10.7554/eLife.29808
Hirsh, J.B., Morisano, D., Peterson, J.B., 2008. Delay discounting: Interactions between personality and cognitive ability. J. Res. Personal. 42, 1646–1650. https://doi.org/10.1016/j.jrp.2008.07.005
Hobson, J.A., McCarley, R.W., Wyzinski, P.W., 1975. Sleep cycle oscillation: reciprocal discharge by two brainstem neuronal groups. Science 189, 55–58. https://doi.org/10.1126/science.1094539
Hoffer, B.J., Siggins, G.R., Oliver, A.P., Bloom, F.E., 1973. Activation of the Pathway from Locus Coeruleus to Rat Cerebellar Purkinje Neurons: Pharmacological Evidence of Noradrenrtgic Central Inhibition. J. Pharmacol. Exp. Ther. 184, 553–569.
Hogarth, R.M., 2005. Deciding analytically or trusting your intuition? The advantages and disadvantages of analytic and intuitive thought, in: The Routines of Decision Making. Lawrence Erlbaum Associates Publishers, Mahwah, NJ, US, pp. 67–82.
Holroyd, C., 2016. The waste disposal problem of effortful control, in: Motivation and Cognitive Control. Routledge, pp. 235–260.
Holroyd, C.B., McClure, S.M., 2015. Hierarchical control over effortful behavior by rodent medial frontal cortex: A computational model. Psychol. Rev. 122, 54–83. https://doi.org/10.1037/a0038339
Hopkins, W.F., Johnston, D., 1988. Noradrenergic enhancement of long-term potentiation at mossy fiber synapses in the hippocampus. J. Neurophysiol. 59, 667–687. https://doi.org/10.1152/jn.1988.59.2.667
Hopkins, W.F., Johnston, D., 1984. Frequency-dependent noradrenergic modulation of long-term potentiation in the hippocampus. Science
References
280
226, 350–352. https://doi.org/10.1126/science.6091272
Hori, Y., Minamimoto, T., Kimura, M., 2009. Neuronal Encoding of Reward Value and Direction of Actions in the Primate Putamen. J. Neurophysiol. 102, 3530–3543. https://doi.org/10.1152/jn.00104.2009
Hosokawa, T., Kennerley, S.W., Sloan, J., Wallis, J.D., 2013. Single-Neuron Mechanisms Underlying Cost-Benefit Analysis in Frontal Cortex. J. Neurosci. 33, 17385–17397. https://doi.org/10.1523/JNEUROSCI.2221-13.2013
Huber, O., 1980. The influence of some task variables on cognitive operations in an information-processing decision model. Acta Psychol. (Amst.) 45, 187–196. https://doi.org/10.1016/0001-6918(80)90031-1
Hui, S., Ghergurovich, J.M., Morscher, R.J., Jang, C., Teng, X., Lu, W., Esparza, L.A., Reya, T., Le Zhan, null, Yanxiang Guo, J., White, E., Rabinowitz, J.D., 2017. Glucose feeds the TCA cycle via circulating lactate. Nature 551, 115–118. https://doi.org/10.1038/nature24057
Hull, C.L., 1943. Principles of behavior: an introduction to behavior theory, Principles of behavior: an introduction to behavior theory. Appleton-Century, Oxford, England.
Hunt, L.T., Kolling, N., Soltani, A., Woolrich, M.W., Rushworth, M.F.S., Behrens, T.E.J., 2012. Mechanisms underlying cortical activity during value-guided choice. Nat. Neurosci. 15, 470–476. https://doi.org/10.1038/nn.3017
Hupé, J.-M., Lamirel, C., Lorenceau, J., 2009. Pupil dynamics during bistable motion perception. J. Vis. 9, 10–10. https://doi.org/10.1167/9.7.10
Hurley, L., Devilbiss, D., Waterhouse, B., 2004. A matter of focus: monoaminergic modulation of stimulus coding in mammalian sensory networks. Curr. Opin. Neurobiol. 14, 488–495. https://doi.org/10.1016/j.conb.2004.06.007
Ikemoto, S., Yang, C., Tan, A., 2015. Basal ganglia circuit loops, dopamine and motivation: A review and enquiry. Behav. Brain Res. 290, 17–31. https://doi.org/10.1016/j.bbr.2015.04.018
Inzlicht, M., Schmeichel, B.J., 2012. What Is Ego Depletion? Toward a Mechanistic Revision of the
Resource Model of Self-Control. Perspect. Psychol. Sci. 7, 450–463. https://doi.org/10.1177/1745691612454134
Inzlicht, M., Schmeichel, B.J., Macrae, C.N., 2014. Why self-control seems (but may not be) limited. Trends Cogn. Sci. 18, 127–133. https://doi.org/10.1016/j.tics.2013.12.009
Ishimatsu, M., Williams, J.T., 1996. Synchronous Activity in Locus Coeruleus Results from Dendritic Interactions in Pericoerulear Regions. J. Neurosci. 16, 5196–5204. https://doi.org/10.1523/JNEUROSCI.16-16-05196.1996
Jahn, A., Nee, D.E., Alexander, W.H., Brown, J.W., 2014. Distinct regions of anterior cingulate cortex signal prediction and outcome evaluation. NeuroImage 95, 80–89. https://doi.org/10.1016/j.neuroimage.2014.03.050
Jahn, C.I., Gilardeau, S., Varazzani, C., Blain, B., Sallet, J., Walton, M.E., Bouret, S., 2018. Dual contributions of noradrenaline to behavioural flexibility and motivation. Psychopharmacology (Berl.) 235, 2687–2702. https://doi.org/10.1007/s00213-018-4963-z
Jahn, C.I., Varazzani, C., Sallet, J., Walton, M.E., Bouret, S., 2020. Noradrenergic But Not Dopaminergic Neurons Signal Task State Changes and Predict Reengagement After a Failure. Cereb. Cortex N. Y. N 1991 30, 4979–4994. https://doi.org/10.1093/cercor/bhaa089
Jamadar, S., Thienel, R., Karayanidis, F., 2015. Task switching processes. Brain Mapp. Encycl. Ref. 327–335. https://doi.org/10.1016/B978-0-12-397025-1.00250-5
Jepma, M., Brown, S.B.R.E., Murphy, P.R., Koelewijn, S.C., de Vries, B., van den Maagdenberg, A.M., Nieuwenhuis, S., 2018. Noradrenergic and Cholinergic Modulation of Belief Updating. J. Cogn. Neurosci. 30, 1803–1820. https://doi.org/10.1162/jocn_a_01317
Jepma, M., Deinum, J., Asplund, C.L., Rombouts, S.A., Tamsma, J.T., Tjeerdema, N., Spapé, M.M., Garland, E.M., Robertson, D., Lenders, J.W., Nieuwenhuis, S., 2011. Neurocognitive function in dopamine-β-hydroxylase deficiency. Neuropsychopharmacol. Off. Publ. Am. Coll. Neuropsychopharmacol. 36, 1608–1619. https://doi.org/10.1038/npp.2011.42
References
281
Jepma, M., Nieuwenhuis, S., 2011. Pupil Diameter Predicts Changes in the Exploration–Exploitation Trade-off: Evidence for the Adaptive Gain Theory. J. Cogn. Neurosci. 23, 1587–1596. https://doi.org/10.1162/jocn.2010.21548
Jett, J.D., Bulin, S.E., Hatherall, L.C., McCartney, C.M., Morilak, D.A., 2017. Deficits in cognitive flexibility induced by chronic unpredictable stress are associated with impaired glutamate neurotransmission in the rat medial prefrontal cortex. Neuroscience 346, 284–297. https://doi.org/10.1016/j.neuroscience.2017.01.017
Jiang, M., Griff, E.R., Ennis, M., Zimmer, L.A., Shipley, M.T., 1996. Activation of Locus Coeruleus Enhances the Responses of Olfactory Bulb Mitral Cells to Weak Olfactory Nerve Input. J. Neurosci. 16, 6319–6329. https://doi.org/10.1523/JNEUROSCI.16-19-06319.1996
Jimura, K., Locke, H.S., Braver, T.S., 2010. Prefrontal cortex mediation of cognitive enhancement in rewarding motivational contexts. Proc. Natl. Acad. Sci. 107, 8871–8876. https://doi.org/10.1073/pnas.1002007107
Job, V., Dweck, C.S., Walton, G.M., 2010. Ego Depletion—Is It All in Your Head?: Implicit Theories About Willpower Affect Self-Regulation. Psychol. Sci. 21, 1686–1693. https://doi.org/10.1177/0956797610384745
Jodo, E., Takeuchi, S., Kayama, Y., 1995. P3b-like potential of rats recorded in an active discrimination task. Electroencephalogr. Clin. Neurophysiol. 96, 555–560. https://doi.org/10.1016/0013-4694(95)00067-9
Johns, M., Inzlicht, M., Schmader, T., 2008. Stereotype threat and executive resource depletion: Examining the influence of emotion regulation. J. Exp. Psychol. Gen. 137, 691–705. https://doi.org/10.1037/a0013834
Johnson, E.J., Payne, J.W., 1985. Effort and Accuracy in Choice. Manag. Sci. 31, 395–414. https://doi.org/10.1287/mnsc.31.4.395
Jorna, P.G.A.M., 1992. Spectral analysis of heart rate and psychological state: A review of its validity as a workload index. Biol. Psychol. 34, 237–257. https://doi.org/10.1016/0301-0511(92)90017-O
Joshi, S., Gold, J.I., 2020. Context-Dependent Relationships between Locus Coeruleus Firing Patterns and Coordinated Neural Activity in the Anterior Cingulate Cortex. bioRxiv 2020.09.26.314831. https://doi.org/10.1101/2020.09.26.314831
Joshi, S., Li, Y., Kalwani, R.M., Gold, J.I., 2016. Relationships between Pupil Diameter and Neuronal Activity in the Locus Coeruleus, Colliculi, and Cingulate Cortex. Neuron 89, 221–234. https://doi.org/10.1016/j.neuron.2015.11.028
Jouvet, M., 1969. Biogenic amines and the states of sleep. Science 163, 32–41. https://doi.org/10.1126/science.163.3862.32
Kable, J.W., Glimcher, P.W., 2007. The neural correlates of subjective value during intertemporal choice. Nat. Neurosci. 10, 1625–1633. https://doi.org/10.1038/nn2007
Kahneman, D., 1973. Attention and effort. Prentice-Hall, Englewood Cliffs, N.J.
Kahneman, D., Beatty, J., 1966. Pupil diameter and load on memory. Science 154, 1583–1585. https://doi.org/10.1126/science.154.3756.1583
Kahneman, D., Frederick, S., 2005. A Model of Heuristic Judgment, in: Holyoak, K., Morrison, B. (Eds.), The Cambridge Handbook of Thinking and Reasoning. Cambridge University Press, pp. 267–293.
Kahneman, D., Tversky, A., 1979. Prospect Theory: An Analysis of Decision under Risk. Econometrica 47, 263–291. https://doi.org/10.2307/1914185
Kalwani, R.M., Joshi, S., Gold, J.I., 2014. Phasic Activation of Individual Neurons in the Locus Ceruleus/Subceruleus Complex of Monkeys Reflects Rewarded Decisions to Go But Not Stop. J. Neurosci. 34, 13656–13669. https://doi.org/10.1523/JNEUROSCI.2566-14.2014
Kamp, S.-M., Donchin, E., 2015. ERP and pupil responses to deviance in an oddball paradigm. Psychophysiology 52, 460–471. https://doi.org/10.1111/psyp.12378
Kane, G.A., Vazey, E.M., Wilson, R.C., Shenhav, A., Daw, N.D., Aston-Jones, G., Cohen, J.D., 2017. Increased locus coeruleus tonic activity causes
References
282
disengagement from a patch-foraging task. Cogn. Affect. Behav. Neurosci. 17, 1073–1083. https://doi.org/10.3758/s13415-017-0531-y
Kawahara, H., Kawahara, Y., Westerink, B.H., 2001. The noradrenaline-dopamine interaction in the rat medial prefrontal cortex studied by multi-probe microdialysis. Eur. J. Pharmacol. 418, 177–186. https://doi.org/10.1016/s0014-2999(01)00863-9
Kebschull, J.M., da Silva, P.G., Reid, A.P., Peikon, I.D., Albeanu, D.F., Zador, A.M., 2016. High-throughput mapping of single neuron projections by sequencing of barcoded RNA. Neuron 91, 975–987. https://doi.org/10.1016/j.neuron.2016.07.036
Kempadoo, K.A., Mosharov, E.V., Choi, S.J., Sulzer, D., Kandel, E.R., 2016. Dopamine release from the locus coeruleus to the dorsal hippocampus promotes spatial learning and memory. Proc. Natl. Acad. Sci. 113, 14835–14840. https://doi.org/10.1073/pnas.1616515114
Kennedy, D.O., Scholey, A.B., 2000. Glucose administration, heart rate and cognitive performance: effects of increasing mental effort. Psychopharmacology (Berl.) 149, 63–71. https://doi.org/10.1007/s002139900335
Kennerley, S.W., Behrens, T.E.J., Wallis, J.D., 2011. Double dissociation of value computations in orbitofrontal and anterior cingulate neurons. Nat. Neurosci. 14, 1581–1589. https://doi.org/10.1038/nn.2961
Kennerley, S.W., Dahmubed, A.F., Lara, A.H., Wallis, J.D., 2009. Neurons in the frontal lobe encode the value of multiple decision variables. J. Cogn. Neurosci. 21, 1162–1178. https://doi.org/10.1162/jocn.2009.21100
Kennerley, S.W., Wallis, J.D., 2009a. Reward-dependent modulation of working memory in lateral prefrontal cortex. J. Neurosci. Off. J. Soc. Neurosci. 29, 3259–3270. https://doi.org/10.1523/JNEUROSCI.5353-08.2009
Kennerley, S.W., Wallis, J.D., 2009b. Evaluating choices by single neurons in the frontal lobe: outcome value encoded across multiple decision variables. Eur. J. Neurosci. 29, 2061–2073. https://doi.org/10.1111/j.1460-9568.2009.06743.x
Kirby, K.N., MarakoviĆ, N.N., 1996. Delay-discounting probabilistic rewards: Rates decrease as amounts increase. Psychon. Bull. Rev. 3, 100–104. https://doi.org/10.3758/BF03210748
Kirchner, W.K., 1958. Age differences in short-term retention of rapidly changing information. J. Exp. Psychol. 55, 352–358. https://doi.org/10.1037/h0043688
Kirino, E., Belger, A., Goldman-Rakic, P., McCarthy, G., 2000. Prefrontal activation evoked by infrequent target and novel stimuli in a visual target detection task: an event-related functional magnetic resonance imaging study. J. Neurosci. Off. J. Soc. Neurosci. 20, 6612–6618.
Kleinsorge, T., Rinkenauer, G., 2012. Effects of monetary incentives on task switching. Exp. Psychol. 59, 216–226. https://doi.org/10.1027/1618-3169/a000146
Klimesch, W., 2011. Evoked alpha and early access to the knowledge system: The P1 inhibition timing hypothesis. Brain Res. 1408, 52–71. https://doi.org/10.1016/j.brainres.2011.06.003
Koga, K., Yamada, A., Song, Q., Li, X.-H., Chen, Q.-Y., Liu, R.-H., Ge, J., Zhan, C., Furue, H., Zhuo, M., Chen, T., 2020. Ascending noradrenergic excitation from the locus coeruleus to the anterior cingulate cortex. Mol. Brain 13, 49. https://doi.org/10.1186/s13041-020-00586-5
Kolling, N., Behrens, T., Wittmann, M., Rushworth, M., 2016. Multiple signals in anterior cingulate cortex. Curr. Opin. Neurobiol. 37, 36–43. https://doi.org/10.1016/j.conb.2015.12.007
Koob, G.F., 1999. Corticotropin-releasing factor, norepinephrine, and stress. Biol. Psychiatry 46, 1167–1180. https://doi.org/10.1016/s0006-3223(99)00164-x
Kool, W., Botvinick, M., 2018. Mental labour. Nat. Hum. Behav. 2, 899–908. https://doi.org/10.1038/s41562-018-0401-9
Kool, W., Botvinick, M., 2014. A labor/leisure tradeoff in cognitive control. J. Exp. Psychol. Gen. 143, 131–141. https://doi.org/10.1037/a0031048
Kool, W., McGuire, J.T., Rosen, Z.B., Botvinick, M.M., 2010. Decision making and the avoidance
References
283
of cognitive demand. J. Exp. Psychol. Gen. 139, 665–682. https://doi.org/10.1037/a0020198
Kopp, B., Steinke, A., Visalli, A., 2020. Cognitive flexibility and N2/P3 event-related brain potentials. Sci. Rep. 10, 9859. https://doi.org/10.1038/s41598-020-66781-5
Kossl, M., Vater, M., 1989. Noradrenaline enhances temporal auditory contrast and neuronal timing precision in the cochlear nucleus of the mustached bat. J. Neurosci. 9, 4169–4178. https://doi.org/10.1523/JNEUROSCI.09-12-04169.1989
Krajbich, I., Armel, C., Rangel, A., 2010. Visual fixations and the computation and comparison of value in simple choice. Nat. Neurosci. 13, 1292–1298. https://doi.org/10.1038/nn.2635
Krajbich, I., Rangel, A., 2011. Multialternative drift-diffusion model predicts the relationship between visual fixations and choice in value-based decisions. Proc. Natl. Acad. Sci. 108, 13852–13857. https://doi.org/10.1073/pnas.1101328108
Krebs, R.M., Boehler, C.N., Woldorff, M.G., 2010. The influence of reward associations on conflict processing in the Stroop task. Cognition 117, 341–347. https://doi.org/10.1016/j.cognition.2010.08.018
Kurniawan, I.T., Guitart-Masip, M., Dayan, P., Dolan, R.J., 2013. Effort and Valuation in the Brain: The Effects of Anticipation and Execution. J. Neurosci. 33, 6160–6169. https://doi.org/10.1523/JNEUROSCI.4777-12.2013
Kurniawan, I.T., Seymour, B., Talmi, D., Yoshida, W., Chater, N., Dolan, R.J., 2010. Choosing to Make an Effort: The Role of Striatum in Signaling Physical Effort of a Chosen Action. J. Neurophysiol. 104, 313–321. https://doi.org/10.1152/jn.00027.2010
Kurth-Nelson, Z., Bickel, W., Redish, A.D., 2012. A theoretical account of cognitive effects in delay discounting. Eur. J. Neurosci. 35, 1052–1064. https://doi.org/10.1111/j.1460-9568.2012.08058.x
Kurzban, R., 2010. Does the Brain Consume Additional Glucose during Self-Control Tasks? Evol. Psychol. 8, 147470491000800200. https://doi.org/10.1177/147470491000800208
Kurzban, R., Duckworth, A., Kable, J.W., Myers, J., 2013. An opportunity cost model of subjective effort and task performance. Behav. Brain Sci. 36. https://doi.org/10.1017/S0140525X12003196
La Camera, G., Bouret, S., Richmond, B.J., 2018. Contributions of Lateral and Orbital Frontal Regions to Abstract Rule Acquisition and Reversal in Monkeys. Front. Neurosci. 12, 165. https://doi.org/10.3389/fnins.2018.00165
Lacey, J., 1967. Somatic response patterning and stress : some revisions of activation theory. Psychol. Stress Issues Res. 14–37.
Lapiz, M.D.S., Bondi, C.O., Morilak, D.A., 2007. Chronic treatment with desipramine improves cognitive performance of rats in an attentional set-shifting test. Neuropsychopharmacol. Off. Publ. Am. Coll. Neuropsychopharmacol. 32, 1000–1010. https://doi.org/10.1038/sj.npp.1301235
Lapiz, M.D.S., Morilak, D.A., 2006. Noradrenergic modulation of cognitive function in rat medial prefrontal cortex as measured by attentional set shifting capability. Neuroscience 137, 1039–1049. https://doi.org/10.1016/j.neuroscience.2005.09.031
Larsen, T.S., Rasmussen, P., Overgaard, M., Secher, N.H., Nielsen, H.B., 2008. Non-selective β-adrenergic blockade prevents reduction of the cerebral metabolic ratio during exhaustive exercise in humans. J. Physiol. 586, 2807–2815. https://doi.org/10.1113/jphysiol.2008.151449
Le Bouc, R., Rigoux, L., Schmidt, L., Degos, B., Welter, M.-L., Vidailhet, M., Daunizeau, J., Pessiglione, M., 2016. Computational Dissection of Dopamine Motor and Motivational Functions in Humans. J. Neurosci. 36, 6623–6633. https://doi.org/10.1523/JNEUROSCI.3078-15.2016
Lecas, J.-C., 2004. Locus coeruleus activation shortens synaptic drive while decreasing spike latency and jitter in sensorimotor cortex. Implications for neuronal integration. Eur. J. Neurosci. 19, 2519–2530. https://doi.org/10.1111/j.0953-816X.2004.03341.x
LeDoux, J.E., 2000. Emotion circuits in the brain. Annu. Rev. Neurosci. 23, 155–184.
References
284
https://doi.org/10.1146/annurev.neuro.23.1.155
Lee, C.-M., Chang, W.-C., Chang, K.-B., Shyu, B.-C., 2007. Synaptic organization and input-specific short-term plasticity in anterior cingulate cortical neurons with intact thalamic inputs. Eur. J. Neurosci. 25, 2847–2861. https://doi.org/10.1111/j.1460-9568.2007.05485.x
Lemon, N., Aydin-Abidin, S., Funke, K., Manahan-Vaughan, D., 2009. Locus Coeruleus Activation Facilitates Memory Encoding and Induces Hippocampal LTD that Depends on β-Adrenergic Receptor Activation. Cereb. Cortex N. Y. NY 19, 2827–2837. https://doi.org/10.1093/cercor/bhp065
Lempert, K.M., Chen, Y.L., Fleming, S.M., 2015. Relating Pupil Dilation and Metacognitive Confidence during Auditory Decision-Making. PloS One 10, e0126588. https://doi.org/10.1371/journal.pone.0126588
Léna, C., de Kerchove D’Exaerde, A., Cordero-Erausquin, M., Le Novère, N., del Mar Arroyo-Jimenez, M., Changeux, J.P., 1999. Diversity and distribution of nicotinic acetylcholine receptors in the locus ceruleus neurons. Proc. Natl. Acad. Sci. U. S. A. 96, 12126–12131. https://doi.org/10.1073/pnas.96.21.12126
Leotti, L.A., Wager, T.D., 2010. Motivational influences on response inhibition measures. J. Exp. Psychol. Hum. Percept. Perform. 36, 430–447. https://doi.org/10.1037/a0016802
Lestienne, R., 2001. Spike timing, synchronization and information processing on the sensory side of the central nervous system. Prog. Neurobiol. 65, 545–591. https://doi.org/10.1016/s0301-0082(01)00019-3
Levitt, P., Moore, R.Y., 1978. Noradrenaline neuron innervation of the neocortex in the rat. Brain Res. 139, 219–231. https://doi.org/10.1016/0006-8993(78)90925-3
Lewis, D.A., Morrison, J.H., 1989. Noradrenergic innervation of monkey prefrontal cortex: a dopamine-beta-hydroxylase immunohistochemical study. J. Comp. Neurol. 282, 317–330. https://doi.org/10.1002/cne.902820302
Li, B.-M., Mao, Z.-M., Wang, M., Mei, Z.-T., 1999. Alpha-2 Adrenergic Modulation of Prefrontal Cortical Neuronal Activity Related to Spatial Working Memory in Monkeys. Neuropsychopharmacology 21, 601–610. https://doi.org/10.1016/S0893-133X(99)00070-6
Li, B.M., Mei, Z.T., 1994. Delayed-response deficit induced by local injection of the alpha 2-adrenergic antagonist yohimbine into the dorsolateral prefrontal cortex in young adult monkeys. Behav. Neural Biol. 62, 134–139. https://doi.org/10.1016/s0163-1047(05)80034-2
Libby, R., Lipe, M.G., 1992. Incentives, Effort, and the Cognitive Processes Involved in Accounting-Related Judgments. J. Account. Res. 30, 249–273. https://doi.org/10.2307/2491126
Lieder, F., Griffiths, T.L., 2017. Strategy selection as rational metareasoning. Psychol. Rev. 124, 762–794. https://doi.org/10.1037/rev0000075
Lieder, F., Griffiths, T.L., Goodman, N.D., 2012. Burn-in, bias, and the rationality of anchoring. Presented at the NIPS.
Lieder, F., Plunkett, D., Hamrick, J.B., Russell, S.J., Hay, N.J., Griffiths, T.L., 2014. Algorithm selection by rational metareasoning as a model of human strategy selection, in: Proceedings of the 27th International Conference on Neural Information Processing Systems - Volume 2, NIPS’14. MIT Press, Cambridge, MA, USA, pp. 2870–2878.
Likhtik, E., Johansen, J.P., 2019. Neuromodulation in circuits of aversive emotional learning. Nat. Neurosci. 22, 1586–1597. https://doi.org/10.1038/s41593-019-0503-3
Liu, G., Feng, D., Wang, J., Zhang, H., Peng, Z., Cai, M., Yang, J., Zhang, R., Wang, H., Wu, S., Tan, Q., 2018. rTMS Ameliorates PTSD Symptoms in Rats by Enhancing Glutamate Transmission and Synaptic Plasticity in the ACC via the PTEN/Akt Signalling Pathway. Mol. Neurobiol. 55, 3946–3958. https://doi.org/10.1007/s12035-017-0602-7
Liu, K.Y., Marijatta, F., Hämmerer, D., Acosta-Cabronero, J., Düzel, E., Howard, R.J., 2017. Magnetic resonance imaging of the human locus coeruleus: A systematic review. Neurosci. Biobehav. Rev. 83, 325–355.
References
285
https://doi.org/10.1016/j.neubiorev.2017.10.023
Lodish, H., Berk, A., Zipursky, S.L., Matsudaira, P., Baltimore, D., Darnell, J., 2000. Chemical Equilibrium. Mol. Cell Biol. 4th Ed.
Logan, G.D., 1985. Skill and automaticity: Relations, implications, and future directions. Can. J. Psychol. Can. Psychol. 39, 367–386. https://doi.org/10.1037/h0080066
LoTemplio, S., Silcox, J., Federmeier, K.D., Payne, B.R., 2021. Inter- and intra-individual coupling between pupillary, electrophysiological, and behavioral responses in a visual oddball task. Psychophysiology 58, e13758. https://doi.org/10.1111/psyp.13758
Loughlin, S.E., Foote, S.L., Grzanna, R., 1986. Efferent projections of nucleus locus coeruleus: Morphologic subpopulations have different efferent targets. Neuroscience 18, 307–319. https://doi.org/10.1016/0306-4522(86)90156-9
Luck, S.J., Vogel, E.K., 1997. The capacity of visual working memory for features and conjunctions. Nature 390, 279–281. https://doi.org/10.1038/36846
Luppi, P.-H., Aston-Jones, G., Akaoka, H., Chouvet, G., Jouvet, M., 1995. Afferent projections to the rat locus coeruleus demonstrated by retrograde and anterograde tracing with cholera-toxin B subunit and Phaseolus vulgaris leucoagglutinin. Neuroscience 65, 119–160. https://doi.org/10.1016/0306-4522(94)00481-J
Luu, P., Tucker, D.M., 2001. Regulating action: alternating activation of midline frontal and motor cortical networks. Clin. Neurophysiol. Off. J. Int. Fed. Clin. Neurophysiol. 112, 1295–1306. https://doi.org/10.1016/s1388-2457(01)00559-4
Ma, C.-L., Qi, X.-L., Peng, J.-Y., Li, B.-M., 2003. Selective deficit in no-go performance induced by blockade of prefrontal cortical alpha 2-adrenoceptors in monkeys. Neuroreport 14, 1013–1016. https://doi.org/10.1097/01.wnr.0000070831.57864.7b
Madden, G.J., Begotka, A.M., Raiff, B.R., Kastern, L.L., 2003. Delay discounting of real and hypothetical rewards. Exp. Clin.
Psychopharmacol. 11, 139–145. https://doi.org/10.1037/1064-1297.11.2.139
Magistretti, P.J., 1988. Regulation of glycogenolysis by neurotransmitters in the central nervous system. Diabete Metab. 14, 237–246.
Magistretti, P.J., Morrison, J.H., 1988. Noradrenaline- and vasoactive intestinal peptide-containing neuronal systems in neocortex: functional convergence with contrasting morphology. Neuroscience 24, 367–378. https://doi.org/10.1016/0306-4522(88)90338-7
Magistretti, P.J., Morrison, J.H., Shoemaker, W.J., Sapin, V., Bloom, F.E., 1981. Vasoactive intestinal polypeptide induces glycogenolysis in mouse cortical slices: a possible regulatory mechanism for the local control of energy metabolism. Proc. Natl. Acad. Sci. U. S. A. 78, 6535–6539. https://doi.org/10.1073/pnas.78.10.6535
Maity, S., Chandanathil, M., Millis, R.M., Connor, S.A., 2020. Norepinephrine stabilizes translation-dependent, homosynaptic long-term potentiation through mechanisms requiring the cAMP sensor Epac, mTOR and MAPK. Eur. J. Neurosci. 52, 3679–3688. https://doi.org/10.1111/ejn.14735
Manalis, R.S., 1977. Voltage-dependent effect of curare at the frog neuromuscular junction. Nature 267, 366–368. https://doi.org/10.1038/267366a0
Manes, F., Sahakian, B., Clark, L., Rogers, R., Antoun, N., Aitken, M., Robbins, T., 2002. Decision-making processes following damage to the prefrontal cortex. Brain J. Neurol. 125, 624–639. https://doi.org/10.1093/brain/awf049
Mäntylä, T., Still, J., Gullberg, S., Del Missier, F., 2012. Decision Making in Adults With ADHD. J. Atten. Disord. 16, 164–173. https://doi.org/10.1177/1087054709360494
Manunta, Y., Edeline, J.-M., 2004. Noradrenergic induction of selective plasticity in the frequency tuning of auditory cortex neurons. J. Neurophysiol. 92, 1445–1463. https://doi.org/10.1152/jn.00079.2004
Mao, Z.-M., Arnsten, A.F.T., Li, B.-M., 1999. Local infusion of an α-1 adrenergic agonist into the prefrontal cortex impairs spatial working
References
286
memory performance in monkeys. Biol. Psychiatry 46, 1259–1265. https://doi.org/10.1016/S0006-3223(99)00139-0
Marcora, S., 2009. Perception of effort during exercise is independent of afferent feedback from skeletal muscles, heart, and lungs. J. Appl. Physiol. 106, 2060–2062. https://doi.org/10.1152/japplphysiol.90378.2008
Mars, R.B., Coles, M.G.H., Grol, M.J., Holroyd, C.B., Nieuwenhuis, S., Hulstijn, W., Toni, I., 2005. Neural dynamics of error processing in medial frontal cortex. NeuroImage 28, 1007–1013. https://doi.org/10.1016/j.neuroimage.2005.06.041
Martin, K., Meeusen, R., Thompson, K.G., Keegan, R., Rattray, B., 2018. Mental Fatigue Impairs Endurance Performance: A Physiological Explanation. Sports Med. 48, 2041–2051. https://doi.org/10.1007/s40279-018-0946-9
Martínez-Horta, S., Riba, J., de Bobadilla, R.F., Pagonabarraga, J., Pascual-Sedano, B., Antonijoan, R.M., Romero, S., Mañanas, M.À., García-Sanchez, C., Kulisevsky, J., 2014. Apathy in Parkinson’s Disease: Neurophysiological Evidence of Impaired Incentive Processing. J. Neurosci. 34, 5918–5926.
Mather, M., Clewett, D., Sakaki, M., Harley, C.W., 2016. Norepinephrine ignites local hotspots of neuronal excitation: How arousal amplifies selectivity in perception and memory. Behav. Brain Sci. 39, e200. https://doi.org/10.1017/S0140525X15000667
Matsui, T., Ishikawa, T., Ito, H., Okamoto, M., Inoue, K., Lee, M., Fujikawa, T., Ichitani, Y., Kawanaka, K., Soya, H., 2012. Brain glycogen supercompensation following exhaustive exercise. J. Physiol. 590, 607–616. https://doi.org/10.1113/jphysiol.2011.217919
Matsui, T., Omuro, H., Liu, Y.-F., Soya, M., Shima, T., McEwen, B.S., Soya, H., 2017. Astrocytic glycogen-derived lactate fuels the brain during exhaustive exercise to maintain endurance capacity. Proc. Natl. Acad. Sci. U. S. A. 114, 6358–6363. https://doi.org/10.1073/pnas.1702739114
Matsui, T., Soya, M., Soya, H., 2019. Endurance and Brain Glycogen: A Clue Toward Understanding Central Fatigue. Adv. Neurobiol.
23, 331–346. https://doi.org/10.1007/978-3-030-27480-1_11
Matsui, T., Soya, S., Okamoto, M., Ichitani, Y., Kawanaka, K., Soya, H., 2011. Brain glycogen decreases during prolonged exercise. J. Physiol. 589, 3383–3393. https://doi.org/10.1113/jphysiol.2011.203570
Matsumoto, N., Minamimoto, T., Graybiel, A.M., Kimura, M., 2001. Neurons in the thalamic CM-Pf complex supply striatal neurons with information about behaviorally significant sensory events. J. Neurophysiol. 85, 960–976. https://doi.org/10.1152/jn.2001.85.2.960
Mayr, U., Keele, S.W., 2000. Changing internal constraints on action: The role of backward inhibition. J. Exp. Psychol. Gen. 129, 4–26. https://doi.org/10.1037/0096-3445.129.1.4
McBride, R.L., Sutin, J., 1976. Projections of the locus coeruleus and adjacent pontine tegmentum in the cat. J. Comp. Neurol. 165, 265–284. https://doi.org/10.1002/cne.901650302
McCall, J.G., Al-Hasani, R., Siuda, E.R., Hong, D.Y., Norris, A.J., Ford, C.P., Bruchas, M.R., 2015. CRH Engagement of the Locus Coeruleus Noradrenergic System Mediates Stress-Induced Anxiety. Neuron 87, 605–620. https://doi.org/10.1016/j.neuron.2015.07.002
McCarthy, G., Luby, M., Gore, J., Goldman-Rakic, P., 1997. Infrequent events transiently activate human prefrontal and parietal cortex as measured by functional MRI. J. Neurophysiol. 77, 1630–1634. https://doi.org/10.1152/jn.1997.77.3.1630
McCarthy, G., Wood, C.C., Williamson, P.D., Spencer, D.D., 1989. Task-dependent field potentials in human hippocampal formation. J. Neurosci. 9, 4253–4268. https://doi.org/10.1523/JNEUROSCI.09-12-04253.1989
McCloskey, D.I., 1981. Centrally-generated commands and cardiovascular control in man. Clin. Exp. Hypertens. 3, 369–378. https://doi.org/10.3109/10641968109033671
McCormick, D.A., 1989. Cholinergic and noradrenergic modulation of thalamocortical processing. Trends Neurosci. 12, 215–221. https://doi.org/10.1016/0166-2236(89)90125-2
References
287
McGaugh, J.L., 2004. The amygdala modulates the consolidation of memories of emotionally arousing experiences. Annu. Rev. Neurosci. 27, 1–28. https://doi.org/10.1146/annurev.neuro.27.070203.144157
McGAUGHY, J., ROSS, R.S., EICHENBAUM, H., 2008. NORADRENERGIC, BUT NOT CHOLINERGIC, DEAFFERENTATION OF PREFRONTAL CORTEX IMPAIRS ATTENTIONAL SET-SHIFTING. Neuroscience 153, 63–71. https://doi.org/10.1016/j.neuroscience.2008.01.064
McGuire, J.T., Botvinick, M.M., 2010. Prefrontal cortex, cognitive control, and the registration of decision costs. Proc. Natl. Acad. Sci. 107, 7922–7926. https://doi.org/10.1073/pnas.0910662107
McKerchar, T.L., Green, L., Myerson, J., Pickford, T.S., Hill, J.C., Stout, S.C., 2009. A comparison of four models of delay discounting in humans. Behav. Processes, Proceedings of the meeting of the Society for the Quantitative Analyses of Behavior (SQAB 2008), which took place in Chicago IL, May 22-24 of 2008 81, 256–259. https://doi.org/10.1016/j.beproc.2008.12.017
McMorris, T., 2020. Cognitive Fatigue Effects on Physical Performance: The Role of Interoception. Sports Med. Auckl. NZ 50, 1703–1708. https://doi.org/10.1007/s40279-020-01320-w
McNamara, C.G., Dupret, D., 2017. Two sources of dopamine for the hippocampus. Trends Neurosci. 40, 383–384. https://doi.org/10.1016/j.tins.2017.05.005
McNay, E.C., Fries, T.M., Gold, P.E., 2000. Decreases in rat extracellular hippocampal glucose concentration associated with cognitive demand during a spatial task. Proc. Natl. Acad. Sci. U. S. A. 97, 2881–2885. https://doi.org/10.1073/pnas.050583697
McNay, E.C., McCarty, R.C., Gold, P.E., 2001. Fluctuations in Brain Glucose Concentration during Behavioral Testing: Dissociations between Brain Areas and between Brain and Blood. Neurobiol. Learn. Mem. 75, 325–337. https://doi.org/10.1006/nlme.2000.3976
Meer, E.V.D., Beyer, R., Horn, J., Foth, M., Bornemann, B., Ries, J., Kramer, J., Warmuth, E., Heekeren, H.R., Wartenburger, I., 2010. Resource allocation and fluid intelligence: Insights from
pupillometry. Psychophysiology 47, 158–169. https://doi.org/10.1111/j.1469-8986.2009.00884.x
Meeusen, R., Duclos, M., Foster, C., Fry, A., Gleeson, M., Nieman, D., Raglin, J., Rietjens, G., Steinacker, J., Urhausen, A., European College of Sport Science, American College of Sports Medicine, 2013. Prevention, diagnosis, and treatment of the overtraining syndrome: joint consensus statement of the European College of Sport Science and the American College of Sports Medicine. Med. Sci. Sports Exerc. 45, 186–205. https://doi.org/10.1249/MSS.0b013e318279a10a
Miller, G.A., 1956. The magical number seven, plus or minus two: some limits on our capacity for processing information. Psychol. Rev. 63, 81–97. https://doi.org/10.1037/h0043158
Miller, H.C., Pattison, K.F., DeWall, C.N., Rayburn-Reeves, R., Zentall, T.R., 2010. Self-Control Without a “Self”?: Common Self-Control Processes in Humans and Dogs. Psychol. Sci. 21, 534–538. https://doi.org/10.1177/0956797610364968
Milosavljevic, M., Malmaud, J., Huth, A., Koch, C., Rangel, A., 2010. The Drift Diffusion Model Can Account for the Accuracy and Reaction Time of Value-Based Choices Under High and Low Time Pressure. SSRN Electron. J. https://doi.org/10.2139/ssrn.1901533
Minamimoto, T., Hori, Y., Kimura, M., 2009a. Roles of the thalamic CM–PF complex—Basal ganglia circuit in externally driven rebias of action. Brain Res. Bull. 78, 75–79. https://doi.org/10.1016/j.brainresbull.2008.08.013
Minamimoto, T., Hori, Y., Kimura, M., 2005. Complementary process to response bias in the centromedian nucleus of the thalamus. Science 308, 1798–1801. https://doi.org/10.1126/science.1109154
Minamimoto, T., Hori, Y., Richmond, B.J., 2012. Is Working More Costly than Waiting in Monkeys? PLOS ONE 7, e48434. https://doi.org/10.1371/journal.pone.0048434
Minamimoto, T., Kimura, M., 2002. Participation of the thalamic CM-Pf complex in attentional orienting. J. Neurophysiol. 87, 3090–3101. https://doi.org/10.1152/jn.2002.87.6.3090
References
288
Minamimoto, T., La Camera, G., Richmond, B.J., 2009b. Measuring and modeling the interaction among reward size, delay to reward, and satiation level on motivation in monkeys. J. Neurophysiol. 101, 437–447. https://doi.org/10.1152/jn.90959.2008
Missier, F.D., Mäntylä, T., Bruin, W.B. de, 2012. Decision-making Competence, Executive Functioning, and General Cognitive Abilities. J. Behav. Decis. Mak. 25, 331–351. https://doi.org/10.1002/bdm.731
Missier, F.D., Mäntylä, T., Bruin, W.B. de, 2010. Executive functions in decision making: An individual differences approach. Think. Reason. 16, 69–97. https://doi.org/10.1080/13546781003630117
Miyake, A., Friedman, N.P., Emerson, M.J., Witzki, A.H., Howerter, A., Wager, T.D., 2000. The unity and diversity of executive functions and their contributions to complex “Frontal Lobe” tasks: a latent variable analysis. Cognit. Psychol. 41, 49–100. https://doi.org/10.1006/cogp.1999.0734
Mobbs, D., Hassabis, D., Seymour, B., Marchant, J.L., Weiskopf, N., Dolan, R.J., Frith, C.D., 2009. Choking on the money: reward-based performance decrements are associated with midbrain activity. Psychol. Sci. 20, 955–962. https://doi.org/10.1111/j.1467-9280.2009.02399.x
Molden, D.C., Hui, C.M., Scholer, A.A., Meier, B.P., Noreen, E.E., D’Agostino, P.R., Martin, V., 2012. Motivational Versus Metabolic Effects of Carbohydrates on Self-Control. Psychol. Sci. 23, 1137–1144. https://doi.org/10.1177/0956797612439069
Monsell, S., 1996. Control of mental processes, in: Unsolved Mysteries of the Mind: Tutorial Essays in Cognition. Erlbaum (Uk) Taylor & Francis, Publ, Oxford, England, pp. 93–148.
Monsell, S., Sumner, P., Waters, H., 2003. Task-set reconfiguration with predictable and unpredictable task switches. Mem. Cognit. 31, 327–342. https://doi.org/10.3758/bf03194391
Morris, R., Bornstein, A., Shenhav, A. (Eds.), 2019. Goal-directed decision making: computations and neural circuits. Academic Press, Elsevier, London San Diego, CA Cambridge.
Morrison, J.H., Foote, S.L., 1986. Noradrenergic and serotoninergic innervation of cortical, thalamic, and tectal visual structures in Old and New World monkeys. J. Comp. Neurol. 243, 117–138. https://doi.org/10.1002/cne.902430110
Morrison, J.H., Grzanna, R., Molliver, M.E., Coyle, J.T., 1978. The distribution and orientation of noradrenergic fibers in neocortex of the rat: An immunofluorescence study. J. Comp. Neurol. 181, 17–39. https://doi.org/10.1002/cne.901810103
Mulder, G., 1986. The Concept and Measurement of Mental Effort, in: Hockey, G.R.J., Gaillard, A.W.K., Coles, M.G.H. (Eds.), Energetics and Human Information Processing, NATO ASI Series. Springer Netherlands, Dordrecht, pp. 175–198. https://doi.org/10.1007/978-94-009-4448-0_12
Mulder, G., 1979. Sinusarrhythmia and Mental Work Load, in: Moray, N. (Ed.), Mental Workload: Its Theory and Measurement, NATO Conference Series. Springer US, Boston, MA, pp. 327–343. https://doi.org/10.1007/978-1-4757-0884-4_19
Mulder, G., Mulder, L.J.M., 1981. Information Processing and Cardiovascular Control. Psychophysiology 18, 392–402. https://doi.org/10.1111/j.1469-8986.1981.tb02470.x
Müller, M.S., 2014. Functional impact of glycogen degradation on astrocytic signalling. Biochem. Soc. Trans. 42, 1311–1315. https://doi.org/10.1042/BST20140157
Muraven, M., Baumeister, R.F., 2000. Self-regulation and depletion of limited resources: Does self-control resemble a muscle? Psychol. Bull. 126, 247–259. https://doi.org/10.1037/0033-2909.126.2.247
Muraven, M., Slessareva, E., 2003. Mechanisms of self-control failure: motivation and limited resources. Pers. Soc. Psychol. Bull. 29, 894–906. https://doi.org/10.1177/0146167203029007008
Muraven, M., Tice, D.M., Baumeister, R.F., 1998. Self-control as a limited resource: Regulatory depletion patterns. J. Pers. Soc. Psychol. 74, 774–789. https://doi.org/10.1037/0022-3514.74.3.774
Murphy, P.R., Moort, M.L. van, Nieuwenhuis, S., 2016. The Pupillary Orienting Response Predicts Adaptive Behavioral Adjustment after Errors.
References
289
PLOS ONE 11, e0151763. https://doi.org/10.1371/journal.pone.0151763
Murphy, P.R., O’Connell, R.G., O’Sullivan, M., Robertson, I.H., Balsters, J.H., 2014. Pupil diameter covaries with BOLD activity in human locus coeruleus: Pupil Diameter and Locus Coeruleus Activity. Hum. Brain Mapp. 35, 4140–4154. https://doi.org/10.1002/hbm.22466
Murphy, P.R., Robertson, I.H., Balsters, J.H., O’connell, R.G., 2011. Pupillometry and P3 index the locus coeruleus–noradrenergic arousal function in humans. Psychophysiology 48, 1532–1543. https://doi.org/10.1111/j.1469-8986.2011.01226.x
Naaijen, J., Lythgoe, D.J., Zwiers, M.P., Hartman, C.A., Hoekstra, P.J., Buitelaar, J.K., Aarts, E., 2018. Anterior cingulate cortex glutamate and its association with striatal functioning during cognitive control. Eur. Neuropsychopharmacol. 28, 381–391. https://doi.org/10.1016/j.euroneuro.2018.01.002
Nagai, T., Satoh, K., Imamoto, K., Maeda, T., 1981. Divergent projections of catecholamine neurons of the locus coeruleus as revealed by fluorescent retrograde double labeling technique. Neurosci. Lett. 23, 117–123. https://doi.org/10.1016/0304-3940(81)90027-6
Nardecchia, F., Orlando, R., Iacovelli, L., Colamartino, M., Fiori, E., Leuzzi, V., Piccinin, S., Nistico, R., Puglisi-Allegra, S., Di Menna, L., Battaglia, G., Nicoletti, F., Pascucci, T., 2018. Targeting mGlu5 Metabotropic Glutamate Receptors in the Treatment of Cognitive Dysfunction in a Mouse Model of Phenylketonuria. Front. Neurosci. 12. https://doi.org/10.3389/fnins.2018.00154
Nassar, M.R., Rumsey, K.M., Wilson, R.C., Parikh, K., Heasly, B., Gold, J.I., 2012. Rational regulation of learning dynamics by pupil-linked arousal systems. Nat. Neurosci. 15, 1040–1046. https://doi.org/10.1038/nn.3130
Navarra, R., Graf, R., Huang, Y., Logue, S., Comery, T., Hughes, Z., Day, M., 2008. Effects of atomoxetine and methylphenidate on attention and impulsivity in the 5-choice serial reaction time test. Prog. Neuropsychopharmacol. Biol. Psychiatry 32, 34–41. https://doi.org/10.1016/j.pnpbp.2007.06.017
Navon, D., 1984. Resources--a theoretical soup stone? Psychol. Rev. 91, 216–234. https://doi.org/10.1037/0033-295X.91.2.216
Nelson, S.R., Schulz, D.W., Passonneau, J.V., Lowry, O.H., 1968. Control of glycogen levels in brain. J. Neurochem. 15, 1271–1279. https://doi.org/10.1111/j.1471-4159.1968.tb05904.x
Neuman, R.S., Harley, C.W., 1983. Long-lasting potentiation of the dentate gyrus population spike by norepinephrine. Brain Res. 273, 162–165. https://doi.org/10.1016/0006-8993(83)91106-x
Newell, A., Simon, H.A., 1972. Human problem solving. Prentice-Hall, Englewood Cliffs, N.J.
Newman, L.A., Darling, J., McGaughy, J., 2008. Atomoxetine reverses attentional deficits produced by noradrenergic deafferentation of medial prefrontal cortex. Psychopharmacology (Berl.) 200, 39–50. https://doi.org/10.1007/s00213-008-1097-8
Nicholas, A.P., Pieribone, V., Hökfelt, T., 1993. Distributions of mRNAs for alpha-2 adrenergic receptor subtypes in rat brain: an in situ hybridization study. J. Comp. Neurol. 328, 575–594. https://doi.org/10.1002/cne.903280409
Nicholson, W., Snyder, C., 2012. Microeconomic theory: basic principles and extensions, 11th ed. Cengage Learning.
Nickel, P., Nachreiner, F., 2003. Sensitivity and Diagnosticity of the 0.1-Hz Component of Heart Rate Variability as an Indicator of Mental Workload. Hum. Factors J. Hum. Factors Ergon. Soc. 45, 575–590. https://doi.org/10.1518/hfes.45.4.575.27094
Nidal, K., Malik, A.S., 2014. EEG/ERP Analysis: Methods and Applications. CRC Press.
Nieuwenhuis, S., 2011. Learning, the P3, and the Locus Coeruleus-Norepinephrine System, Neural Basis of Motivational and Cognitive Control. The MIT Press.
Nieuwenhuis, S., Aston-Jones, G., Cohen, J.D., 2005. Decision making, the P3, and the locus coeruleus-norepinephrine system. Psychol. Bull. 131, 510–532. https://doi.org/10.1037/0033-2909.131.4.510
References
290
Nieuwenhuis, S., de Geus, E.J., Aston-Jones, G., 2011. The anatomical and functional relationship between the P3 and autonomic components of the orienting response. Psychophysiology 48, 162–175. https://doi.org/10.1111/j.1469-8986.2010.01057.x
Nieuwenhuis, S., Monsell, S., 2002. Residual costs in task switching: testing the failure-to-engage hypothesis. Psychon. Bull. Rev. 9, 86–92. https://doi.org/10.3758/bf03196259
Nieuwenhuys, R., Voogd, J., Huijzen, C. van, 2008. The Human Central Nervous System: A Synopsis and Atlas, 4th ed. Steinkopff-Verlag Heidelberg. https://doi.org/10.1007/978-3-540-34686-9
Nir, Y., Mukamel, R., Dinstein, I., Privman, E., Harel, M., Fisch, L., Gelbard-Sagiv, H., Kipervasser, S., Andelman, F., Neufeld, M.Y., Kramer, U., Arieli, A., Fried, I., Malach, R., 2008. Interhemispheric correlations of slow spontaneous neuronal fluctuations revealed in human sensory cortex. Nat. Neurosci. 11, 1100–1108. https://doi.org/10.1038/nn.2177
Norman, D.A., Bobrow, D.G., 1975. On data-limited and resource-limited processes. Cognit. Psychol. 7, 44–64. https://doi.org/10.1016/0010-0285(75)90004-3
Obel, L.F., Müller, M.S., Walls, A.B., Sickmann, H.M., Bak, L.K., Waagepetersen, H.S., Schousboe, A., 2012. Brain glycogen-new perspectives on its metabolic function and regulation at the subcellular level. Front. Neuroenergetics 4, 3. https://doi.org/10.3389/fnene.2012.00003
O’Connor, T.A., Starr, A., 1985. Intracranial potentials correlated with an event-related potential, P300, in the cat. Brain Res. 339, 27–38. https://doi.org/10.1016/0006-8993(85)90618-3
O’Dowd, B.S., Barrington, J., Ng, K.T., Hertz, E., Hertz, L., 1995. Glycogenolytic response of primary chick and mouse cultures of astrocytes to noradrenaline across development. Brain Res. Dev. Brain Res. 88, 220–223. https://doi.org/10.1016/0165-3806(95)00084-q
Olivers, C.N.L., Peters, J., Houtkamp, R., Roelfsema, P.R., 2011. Different states in visual working memory: when it guides attention and when it does not. Trends Cogn. Sci. 15, 327–334. https://doi.org/10.1016/j.tics.2011.05.004
Olson, V.G., Rockett, H.R., Reh, R.K., Redila, V.A., Tran, P.M., Venkov, H.A., Defino, M.C., Hague, C., Peskind, E.R., Szot, P., Raskind, M.A., 2011. The role of norepinephrine in differential response to stress in an animal model of posttraumatic stress disorder. Biol. Psychiatry 70, 441–448. https://doi.org/10.1016/j.biopsych.2010.11.029
Olszewski, J., Baxter, D., 1954. Cytoarchitecture of the human brainstem. By Jerzy Olszewski and Donald Baxter. Published and distributed in North America for S. Karger by J. B. Lippincott Company, Philadelphia and Montreal. 1954. 199 pages. Price $16.00 (Reviewed by Gerhardt von Bonin). J. Comp. Neurol. 101, 825–825. https://doi.org/10.1002/cne.901010308
O’Reilly, J.X., Schüffelgen, U., Cuell, S.F., Behrens, T.E.J., Mars, R.B., Rushworth, M.F.S., 2013. Dissociable effects of surprise and model update in parietal and anterior cingulate cortex. Proc. Natl. Acad. Sci. U. S. A. 110, E3660–E3669. https://doi.org/10.1073/pnas.1305373110
O’Reilly, R.C., Frank, M.J., 2006. Making working memory work: a computational model of learning in the prefrontal cortex and basal ganglia. Neural Comput. 18, 283–328. https://doi.org/10.1162/089976606775093909
Overtoom, C.C.E., Verbaten, M.N., Kemner, C., Kenemans, J.L., van Engeland, H., Buitelaar, J.K., van der Molen, M.W., van der Gugten, J., Westenberg, H., Maes, R. a. A., Koelega, H.S., 2003. Effects of methylphenidate, desipramine, and L-dopa on attention and inhibition in children with Attention Deficit Hyperactivity Disorder. Behav. Brain Res. 145, 7–15. https://doi.org/10.1016/s0166-4328(03)00097-4
Owen, L., Scholey, A.B., Finnegan, Y., Hu, H., Sünram-Lea, S.I., 2012. The effect of glucose dose and fasting interval on cognitive function: a double-blind, placebo-controlled, six-way crossover study. Psychopharmacology (Berl.) 220, 577–589. https://doi.org/10.1007/s00213-011-2510-2
Padmala, S., Pessoa, L., 2011. Reward reduces conflict by enhancing attentional control and biasing visual cortical processing. J. Cogn. Neurosci. 23, 3419–3432. https://doi.org/10.1162/jocn_a_00011
Padmala, S., Pessoa, L., 2010. Interactions between cognition and motivation during response inhibition. Neuropsychologia 48, 558–
References
291
565. https://doi.org/10.1016/j.neuropsychologia.2009.10.017
Page, M.E., Abercrombie, E.D., 1999. Discrete local application of corticotropin-releasing factor increases locus coeruleus discharge and extracellular norepinephrine in rat hippocampus. Synap. N. Y. N 33, 304–313. https://doi.org/10.1002/(SICI)1098-2396(19990915)33:4<304::AID-SYN7>3.0.CO;2-Z
Pageaux, B., Lepers, R., Dietz, K.C., Marcora, S.M., 2014. Response inhibition impairs subsequent self-paced endurance performance. Eur. J. Appl. Physiol. 114, 1095–1105. https://doi.org/10.1007/s00421-014-2838-5
Pattij, T., Vanderschuren, L.J.M.J., 2008. The neuropharmacology of impulsive behaviour. Trends Pharmacol. Sci. 29, 192–199. https://doi.org/10.1016/j.tips.2008.01.002
Payne, J.W., Bettman, J.R., Johnson, E.J., 1988. Adaptive strategy selection in decision making. J. Exp. Psychol. Learn. Mem. Cogn. 14, 534–552. https://doi.org/10.1037/0278-7393.14.3.534
Payzan-LeNestour, E., Dunne, S., Bossaerts, P., O’Doherty, J.P., 2013. The Neural Representation of Unexpected Uncertainty during Value-Based Decision Making. Neuron 79, 191–201. https://doi.org/10.1016/j.neuron.2013.04.037
Pellerin, L., Magistretti, P.J., 1994. Glutamate uptake into astrocytes stimulates aerobic glycolysis: a mechanism coupling neuronal activity to glucose utilization. Proc. Natl. Acad. Sci. U. S. A. 91, 10625–10629. https://doi.org/10.1073/pnas.91.22.10625
Pennington, B.F., Ozonoff, S., 1996. Executive functions and developmental psychopathology. Child Psychol. Psychiatry Allied Discip. 37, 51–87. https://doi.org/10.1111/j.1469-7610.1996.tb01380.x
Perez, D.M., 2020. α1-Adrenergic Receptors in Neurotransmission, Synaptic Plasticity, and Cognition. Front. Pharmacol. 11, 581098. https://doi.org/10.3389/fphar.2020.581098
Pessiglione, M., Vinckier, F., Bouret, S., Daunizeau, J., Le Bouc, R., 2018. Why not try harder? Computational approach to motivation deficits in neuro-psychiatric diseases. Brain 141, 629–650. https://doi.org/10.1093/brain/awx278
Peters, J., Büchel, C., 2009. Overlapping and Distinct Neural Systems Code for Subjective Value during Intertemporal and Risky Decision Making. J. Neurosci. 29, 15727–15734.
Petersen, S.E., Posner, M.I., 2012. The Attention System of the Human Brain: 20 Years After. Annu. Rev. Neurosci. 35, 73–89. https://doi.org/10.1146/annurev-neuro-062111-150525
Pettigrew, J.D., 1982. Pharmacologic control of cortical plasticity. Retina Phila. Pa 2, 360–372.
Pettigrew, J.D., Kasamatsu, T., 1978. Local perfusion of noradrenaline maintains visual cortical plasticity. Nature 271, 761–763. https://doi.org/10.1038/271761a0
Phelps, C.H., 1972. Barbiturate-induced glycogen accumulation in brain. An electron microscopic study. Brain Res. 39, 225–234. https://doi.org/10.1016/0006-8993(72)90797-4
Philiastides, M.G., Ratcliff, R., 2013. Influence of Branding on Preference-Based Decision Making. Psychol. Sci. 24, 1208–1215. https://doi.org/10.1177/0956797612470701
Phillips, M.A., Szabadi, E., Bradshaw, C.M., 2000. Comparison of the effects of clonidine and yohimbine on spontaneous pupillary fluctuations in healthy human volunteers. Psychopharmacology (Berl.) 150, 85–89. https://doi.org/10.1007/s002130000398
Phillips, P.E.M., Walton, M.E., Jhou, T.C., 2007. Calculating utility: preclinical evidence for cost-benefit analysis by mesolimbic dopamine. Psychopharmacology (Berl.) 191, 483–495. https://doi.org/10.1007/s00213-006-0626-6
Pineda, J.A., Foote, S.L., Neville, H.J., 1989. Effects of locus coeruleus lesions on auditory, long-latency, event-related potentials in monkey. J. Neurosci. Off. J. Soc. Neurosci. 9, 81–93.
Plassmann, H., O’Doherty, J.P., Rangel, A., 2010. Appetitive and Aversive Goal Values Are Encoded in the Medial Orbitofrontal Cortex at the Time of Decision Making. J. Neurosci. 30, 10799–10808.
Plewnia, C., Hoppe, J., Cohen, L.G., Gerloff, C., 2004. Improved motor skill acquisition after selective stimulation of central norepinephrine. Neurology 62, 2124–2126.
References
292
https://doi.org/10.1212/01.WNL.0000128041.92710.17
Polich, J., 2007. Updating P300: an integrative theory of P3a and P3b. Clin. Neurophysiol. Off. J. Int. Fed. Clin. Neurophysiol. 118, 2128–2148. https://doi.org/10.1016/j.clinph.2007.04.019
Polyn, S.M., Norman, K.A., Kahana, M.J., 2009. A context maintenance and retrieval model of organizational processes in free recall. Psychol. Rev. 116, 129–156. https://doi.org/10.1037/a0014420
Porat, O., Hassin-Baer, S., Cohen, O.S., Markus, A., Tomer, R., 2014. Asymmetric dopamine loss differentially affects effort to maximize gain or minimize loss. Cortex 51, 82–91. https://doi.org/10.1016/j.cortex.2013.10.004
Posner, M.I., Snyder, C.R.R., 2004. Attention and Cognitive Control, Cognitive psychology: Key readings. Psychology Press, New York, NY, US.
Pratt, N., Willoughby, A., Swick, D., 2011. Effects of working memory load on visual selective attention: behavioral and electrophysiological evidence. Front. Hum. Neurosci. 5, 57. https://doi.org/10.3389/fnhum.2011.00057
Prévost, C., Pessiglione, M., Météreau, E., Cléry-Melin, M.-L., Dreher, J.-C., 2010. Separate Valuation Subsystems for Delay and Effort Decision Costs. J. Neurosci. 30, 14080–14090. https://doi.org/10.1523/JNEUROSCI.2752-10.2010
Proske, U., 2005. What is the role of muscle receptors in proprioception? Muscle Nerve 31, 780–787. https://doi.org/10.1002/mus.20330
Puumala, T., Greijus, S., Narinen, K., Haapalinna, A., Riekkinen, P., Sirviö, J., 1998. Stimulation of alpha-1 adrenergic receptors facilitates spatial learning in rats. Eur. Neuropsychopharmacol. J. Eur. Coll. Neuropsychopharmacol. 8, 17–26. https://doi.org/10.1016/s0924-977x(97)00040-0
Qi, S., Zeng, Q., Luo, Y., Duan, H., Ding, C., Hu, W., Li, H., 2014. Impact of working memory load on cognitive control in trait anxiety: an ERP study. PloS One 9, e111791. https://doi.org/10.1371/journal.pone.0111791
Quach, T.T., Rose, C., Schwartz, J.C., 1978. [3H]Glycogen hydrolysis in brain slices: responses to neurotransmitters and modulation of
noradrenaline receptors. J. Neurochem. 30, 1335–1341. https://doi.org/10.1111/j.1471-4159.1978.tb10464.x
Quilodran, R., Rothé, M., Procyk, E., 2008. Behavioral Shifts and Action Valuation in the Anterior Cingulate Cortex. Neuron 57, 314–325. https://doi.org/10.1016/j.neuron.2007.11.031
Rabbitt, P., 1979. How old and young subjects monitor and control responses for accuracy and speed. Br. J. Psychol. 70, 305–311. https://doi.org/10.1111/j.2044-8295.1979.tb01687.x
Rachlin, H., Jones, B.A., 2008. Social discounting and delay discounting. J. Behav. Decis. Mak. 21, 29–43. https://doi.org/10.1002/bdm.567
Radel, R., Gruet, M., Barzykowski, K., 2019. Testing the ego-depletion effect in optimized conditions. PloS One 14, e0213026. https://doi.org/10.1371/journal.pone.0213026
Rajkowski, J., Kubiak, P., Aston-Jones, G., 1994. Locus coeruleus activity in monkey: Phasic and tonic changes are associated with altered vigilance. Brain Res. Bull. 35, 607–616. https://doi.org/10.1016/0361-9230(94)90175-9
Rajkowski, J., Majczynski, H., Clayton, E., Aston-Jones, G., 2004. Activation of Monkey Locus Coeruleus Neurons Varies With Difficulty and Performance in a Target Detection Task. J. Neurophysiol. 92, 361–371. https://doi.org/10.1152/jn.00673.2003
Rämä, P., Linnankoski, I., Tanila, H., Pertovaara, A., Carlson, S., 1996. Medetomidine, atipamezole, and guanfacine in delayed response performance of aged monkeys. Pharmacol. Biochem. Behav. 55, 415–422. https://doi.org/10.1016/s0091-3057(96)00111-6
Raskind, M.A., Peterson, K., Williams, T., Hoff, D.J., Hart, K., Holmes, H., Homas, D., Hill, J., Daniels, C., Calohan, J., Millard, S.P., Rohde, K., O’Connell, J., Pritzl, D., Feiszli, K., Petrie, E.C., Gross, C., Mayer, C.L., Freed, M.C., Engel, C., Peskind, E.R., 2013. A trial of prazosin for combat trauma PTSD with nightmares in active-duty soldiers returned from Iraq and Afghanistan. Am. J. Psychiatry 170, 1003–1010. https://doi.org/10.1176/appi.ajp.2013.12081133
References
293
Ratcliff, R., McKoon, G., 2008. The diffusion decision model: theory and data for two-choice decision tasks. Neural Comput. 20, 873–922. https://doi.org/10.1162/neco.2008.12-06-420
Rauch, S.L., Whalen, P.J., Shin, L.M., McInerney, S.C., Macklin, M.L., Lasko, N.B., Orr, S.P., Pitman, R.K., 2000. Exaggerated amygdala response to masked facial stimuli in posttraumatic stress disorder: a functional MRI study. Biol. Psychiatry 47, 769–776. https://doi.org/10.1016/s0006-3223(00)00828-3
Renart, A., Machens, C.K., 2014. Variability in neural activity and behavior. Curr. Opin. Neurobiol., Theoretical and computational neuroscience 25, 211–220. https://doi.org/10.1016/j.conb.2014.02.013
Reyes, B. a. S., Drolet, G., Van Bockstaele, E.J., 2008. Dynorphin and stress-related peptides in rat locus coeruleus: contribution of amygdalar efferents. J. Comp. Neurol. 508, 663–675. https://doi.org/10.1002/cne.21683
Reynaud, A.J., Froesel, M., Guedj, C., Ben Hadj Hassen, S., Cléry, J., Meunier, M., Ben Hamed, S., Hadj-Bouziane, F., 2019. Atomoxetine improves attentional orienting in a predictive context. Neuropharmacology 150, 59–69. https://doi.org/10.1016/j.neuropharm.2019.03.012
Ridderinkhof, K.R., Ullsperger, M., Crone, E.A., Nieuwenhuis, S., 2004. The Role of the Medial Frontal Cortex in Cognitive Control. Science 306, 443–447. https://doi.org/10.1126/science.1100301
Robbins, T.W., 1984. Cortical noradrenaline, attention and arousal. Psychol. Med. 14, 13–21. https://doi.org/10.1017/s0033291700003032
Robbins, T.W., Arnsten, A.F.T., 2009. The Neuropsychopharmacology of Fronto-Executive Function: Monoaminergic Modulation. Annu. Rev. Neurosci. 32, 267–287. https://doi.org/10.1146/annurev.neuro.051508.135535
Robinson, E.S.J., 2012. Blockade of noradrenaline re-uptake sites improves accuracy and impulse control in rats performing a five-choice serial reaction time tasks. Psychopharmacology (Berl.) 219, 303–312. https://doi.org/10.1007/s00213-011-2420-3
Robinson, E.S.J., Eagle, D.M., Mar, A.C., Bari, A., Banerjee, G., Jiang, X., Dalley, J.W., Robbins, T.W., 2008. Similar effects of the selective noradrenaline reuptake inhibitor atomoxetine on three distinct forms of impulsivity in the rat. Neuropsychopharmacol. Off. Publ. Am. Coll. Neuropsychopharmacol. 33, 1028–1037. https://doi.org/10.1038/sj.npp.1301487
Rogawski, M.A., Aghajanian, G.K., 1980. Modulation of lateral geniculate neurone excitability by noradrenaline microiontophoresis or locus coeruleus stimulation. Nature 287, 731–734. https://doi.org/10.1038/287731a0
Room, P., Postema, F., Korf, J., 1981. Divergent axon collaterals of rat locus coeruleus neurons: demonstration by a fluorescent double labeling technique. Brain Res. 221, 219–230. https://doi.org/10.1016/0006-8993(81)90773-3
Roullet, P., Sara, S., 1998. Consolidation of memory after its reactivation: involvement of beta noradrenergic receptors in the late phase. Neural Plast. 6, 63–68. https://doi.org/10.1155/NP.1998.63
Rudebeck, P.H., Walton, M.E., Smyth, A.N., Bannerman, D.M., Rushworth, M.F.S., 2006. Separate neural pathways process different decision costs. Nat. Neurosci. 9, 1161–1168. https://doi.org/10.1038/nn1756
Rushworth, M.F.S., Walton, M.E., Kennerley, S.W., Bannerman, D.M., 2004. Action sets and decisions in the medial frontal cortex. Trends Cogn. Sci. 8, 410–417. https://doi.org/10.1016/j.tics.2004.07.009
Rydval, O., Ortmann, A., 2004. How financial incentives and cognitive abilities affect task performance in laboratory settings: An illustration. Econ. Lett. 85, 315–320. https://doi.org/10.1016/j.econlet.2004.04.020
Salamone, J.D., Correa, M., Farrar, A., Mingote, S.M., 2007. Effort-related functions of nucleus accumbens dopamine and associated forebrain circuits. Psychopharmacology (Berl.) 191, 461–482. https://doi.org/10.1007/s00213-006-0668-9
Salgado, H., Köhr, G., Treviño, M., 2012. Noradrenergic ‘Tone’ Determines Dichotomous Control of Cortical Spike-Timing-Dependent Plasticity. Sci. Rep. 2. https://doi.org/10.1038/srep00417
References
294
Samanez-Larkin, G.R., Hollon, N.G., Carstensen, L.L., Knutson, B., 2008. Individual Differences in Insular Sensitivity During Loss Anticipation Predict Avoidance Learning. Psychol. Sci. 19, 320–323. https://doi.org/10.1111/j.1467-9280.2008.02087.x
Samuels, J.A., Whitecotton, S.M., 2011. An effort based analysis of the paradoxical effects of incentives on decision-aided performance. J. Behav. Decis. Mak. 24, 345–360. https://doi.org/10.1002/bdm.691
Sanders, A.F., 1979. Some Remarks on Mental Load, in: Moray, N. (Ed.), Mental Workload: Its Theory and Measurement, NATO Conference Series. Springer US, Boston, MA, pp. 41–77. https://doi.org/10.1007/978-1-4757-0884-4_5
San-Galli, A., Varazzani, C., Abitbol, R., Pessiglione, M., Bouret, S., 2018. Primate Ventromedial Prefrontal Cortex Neurons Continuously Encode the Willingness to Engage in Reward-Directed Behavior. Cereb. Cortex 28, 73–89. https://doi.org/10.1093/cercor/bhw351
Saper, C.B., Scammell, T.E., Lu, J., 2005. Hypothalamic regulation of sleep and circadian rhythms. Nature 437, 1257–1263. https://doi.org/10.1038/nature04284
Sara, S.J., 2009. The locus coeruleus and noradrenergic modulation of cognition. Nat. Rev. Neurosci. 10, 211–223. https://doi.org/10.1038/nrn2573
Sara, S.J., 1985. Noradrenergic Modulation of Selective Attention: Its Role in Memory Retrieval. Ann. N. Y. Acad. Sci. 444, 178–193. https://doi.org/10.1111/j.1749-6632.1985.tb37588.x
Sara, S.J., Bouret, S., 2012. Orienting and reorienting: the locus coeruleus mediates cognition through arousal. Neuron 76, 130–141. https://doi.org/10.1016/j.neuron.2012.09.011
Sasse, L.K., Peters, J., Brassen, S., 2017. Cognitive Control Modulates Effects of Episodic Simulation on Delay Discounting in Aging. Front. Aging Neurosci. 9. https://doi.org/10.3389/fnagi.2017.00058
Schiphof-Godart, L., Roelands, B., Hettinga, F.J., 2018. Drive in Sports: How Mental Fatigue Affects Endurance Performance. Front. Psychol. 9. https://doi.org/10.3389/fpsyg.2018.01383
Schmeichel, B.J., 2007. Attention control, memory updating, and emotion regulation temporarily reduce the capacity for executive control. J. Exp. Psychol. Gen. 136, 241–255. https://doi.org/10.1037/0096-3445.136.2.241
Schouppe, N., Demanet, J., Boehler, C.N., Ridderinkhof, K.R., Notebaert, W., 2014. The role of the striatum in effort-based decision-making in the absence of reward. J. Neurosci. Off. J. Soc. Neurosci. 34, 2148–2154. https://doi.org/10.1523/JNEUROSCI.1214-13.2014
Schwarz, L.A., Luo, L., 2015. Organization of the Locus Coeruleus-Norepinephrine System. Curr. Biol. 25, R1051–R1056. https://doi.org/10.1016/j.cub.2015.09.039
Schwarz, L.A., Miyamichi, K., Gao, X.J., Beier, K.T., Weissbourd, B., DeLoach, K.E., Ren, J., Ibanes, S., Malenka, R.C., Kremer, E.J., Luo, L., 2015. Viral-genetic tracing of the input-output organization of a central noradrenaline circuit. Nature 524, 88–92. https://doi.org/10.1038/nature14600
Schweimer, J., Saft, S., Hauber, W., 2005. Involvement of catecholamine neurotransmission in the rat anterior cingulate in effort-related decision making. Behav. Neurosci. 119, 1687–1692. https://doi.org/10.1037/0735-7044.119.6.1687
Secher, N.H., Seifert, T., Van Lieshout, J.J., 2008. Cerebral blood flow and metabolism during exercise: implications for fatigue. J. Appl. Physiol. 104, 306–314. https://doi.org/10.1152/japplphysiol.00853.2007
Seeley, W.W., Menon, V., Schatzberg, A.F., Keller, J., Glover, G.H., Kenna, H., Reiss, A.L., Greicius, M.D., 2007. Dissociable Intrinsic Connectivity Networks for Salience Processing and Executive Control. J. Neurosci. 27, 2349–2356. https://doi.org/10.1523/JNEUROSCI.5587-06.2007
Segal, D.S., Mandell, A.J., 1970. Behavioral activation of rats during intraventricular infusion of norepinephrine. Proc. Natl. Acad. Sci. U. S. A. 66, 289–293. https://doi.org/10.1073/pnas.66.2.289
Segal, M., Bloom, F.E., 1976. The action of norepinephrine in the rat hippocampus. IV. The effects of locus coeruleus stimulation on evoked
References
295
hippocampal unit activity. Brain Res. 107, 513–525. https://doi.org/10.1016/0006-8993(76)90141-4
Segal, M., Bloom, F.E., 1974. The action of norepinephrine in the rat hippocampus. I. Iontophoretic studies. Brain Res. 72, 79–97. https://doi.org/10.1016/0006-8993(74)90652-0
Sellaro, R., van Leusden, J.W.R., Tona, K.-D., Verkuil, B., Nieuwenhuis, S., Colzato, L.S., 2015. Transcutaneous Vagus Nerve Stimulation Enhances Post-error Slowing. J. Cogn. Neurosci. 27, 2126–2132. https://doi.org/10.1162/jocn_a_00851
Sergeant, J.A., Geurts, H., Oosterlaan, J., 2002. How specific is a deficit of executive functioning for attention-deficit/hyperactivity disorder? Behav. Brain Res. 130, 3–28. https://doi.org/10.1016/s0166-4328(01)00430-2
Servan-Schreiber, D., Printz, H., Cohen, J.D., 1990. A network model of catecholamine effects: gain, signal-to-noise ratio, and behavior. Science 249, 892–895. https://doi.org/10.1126/science.2392679
Shamosh, N.A., DeYoung, C.G., Green, A.E., Reis, D.L., Johnson, M.R., Conway, A.R.A., Engle, R.W., Braver, T.S., Gray, J.R., 2008. Individual Differences in Delay Discounting: Relation to Intelligence, Working Memory, and Anterior Prefrontal Cortex. Psychol. Sci. 19, 904–911. https://doi.org/10.1111/j.1467-9280.2008.02175.x
Shamosh, N.A., Gray, J.R., 2008. Delay discounting and intelligence: A meta-analysis. Intelligence 36, 289–305. https://doi.org/10.1016/j.intell.2007.09.004
Sharma, Y., Xu, T., Graf, W.M., Fobbs, A., Sherwood, C.C., Hof, P.R., Allman, J.M., Manaye, K.F., 2010. Comparative Anatomy of the Locus Coeruleus in Humans and Non-Human Primates. J. Comp. Neurol. 518, 963. https://doi.org/10.1002/cne.22249
Shea-Brown, E., Gilzenrat, M.S., Cohen, J.D., 2008. Optimization of decision making in multilayer networks: the role of locus coeruleus. Neural Comput. 20, 2863–2894. https://doi.org/10.1162/neco.2008.03-07-487
Shekhar, A., Sajdyk, T.S., Keim, S.R., Yoder, K.K., Sanders, S.K., 1999. Role of the basolateral
amygdala in panic disorder. Ann. N. Y. Acad. Sci. 877, 747–750. https://doi.org/10.1111/j.1749-6632.1999.tb09315.x
Shenhav, A., Botvinick, M.M., Cohen, J.D., 2013. The Expected Value of Control: An Integrative Theory of Anterior Cingulate Cortex Function. Neuron 79, 217–240. https://doi.org/10.1016/j.neuron.2013.07.007
Shenhav, A., Musslick, S., Lieder, F., Kool, W., Griffiths, T.L., Cohen, J.D., Botvinick, M.M., 2017. Toward a Rational and Mechanistic Account of Mental Effort. Annu. Rev. Neurosci. 40, 99–124. https://doi.org/10.1146/annurev-neuro-072116-031526
Shenhav, A., Straccia, M.A., Cohen, J.D., Botvinick, M.M., 2014. Anterior cingulate engagement in a foraging context reflects choice difficulty, not foraging value. Nat. Neurosci. 17, 1249–1254. https://doi.org/10.1038/nn.3771
Shidara, M., Mizuhiki, T., Richmond, B.J., 2005. Neuronal firing in anterior cingulate neurons changes modes across trials in single states of multitrial reward schedules. Exp. Brain Res. 163, 242–245. https://doi.org/10.1007/s00221-005-2232-y
Shiffrin, R.M., Schneider, W., 1977. Controlled and automatic human information processing: II. Perceptual learning, automatic attending and a general theory. Psychol. Rev. 84, 127–190. https://doi.org/10.1037/0033-295X.84.2.127
Shiv, B., Fedorikhin, A., 1999. Heart and mind in conflict: The interplay of affect and cognition in consumer decision making. J. Consum. Res. 26, 278–292. https://doi.org/10.1086/209563
Shugan, S.M., 1980. The Cost Of Thinking. J. Consum. Res. 7, 99–111. https://doi.org/10.1086/208799
Shulman, R.G., Hyder, F., Rothman, D.L., 2001. Cerebral energetics and the glycogen shunt: neurochemical basis of functional imaging. Proc. Natl. Acad. Sci. U. S. A. 98, 6417–6422. https://doi.org/10.1073/pnas.101129298
Shyu, B.-C., Vogt, B.A., 2009. Short-term synaptic plasticity in the nociceptive thalamic-anterior cingulate pathway. Mol. Pain 5, 51. https://doi.org/10.1186/1744-8069-5-51
References
296
Sidarus, N., Palminteri, S., Chambon, V., 2019. Cost-benefit trade-offs in decision-making and learning. PLOS Comput. Biol. 15, e1007326. https://doi.org/10.1371/journal.pcbi.1007326
Silvetti, M., Vassena, E., Abrahamse, E., Verguts, T., 2018. Dorsal anterior cingulate-brainstem ensemble as a reinforcement meta-learner. PLOS Comput. Biol. 14, e1006370. https://doi.org/10.1371/journal.pcbi.1006370
Simon, H.A., 1997. Models of Bounded Rationality: Empirically grounded economic reason. MIT Press.
Simon, H.A., 1956. Rational choice and the structure of the environment. Psychol. Rev. 63, 129–138. https://doi.org/10.1037/h0042769
Simon, H.A., 1955. A behavioral model of rational choice. Q. J. Econ. 69, 99–118. https://doi.org/10.2307/1884852
Simpson, H.M., Hale, S.M., 1969. Pupillary changes during a decision-making task. Percept. Mot. Skills 29, 495–498. https://doi.org/10.2466/pms.1969.29.2.495
Simpson, K.L., Waterhouse, B.D., Lin, R.C., 1999. Origin, distribution, and morphology of galaninergic fibers in the rodent trigeminal system. J. Comp. Neurol. 411, 524–534. https://doi.org/10.1002/(sici)1096-9861(19990830)411:3<524::aid-cne13>3.0.co;2-x
Sims, J., Carroll, D., 1990. Cardiovascular and Metabolic Activity at Rest and During Psychological and Physical Challenge in Normotensives and Subjects With Mildly Elevated Blood Pressure. Psychophysiology 27, 149–156. https://doi.org/10.1111/j.1469-8986.1990.tb00366.x
Sims, J., Carroll, D., Turner, J.R., Hewitt, J.K., 1988. Cardiac and Metabolic Activity in Mild Hypertensive and Normotensive Subjects. Psychophysiology 25, 172–178. https://doi.org/10.1111/j.1469-8986.1988.tb00982.x
Sirviö, J., Riekkinen, P., Vajanto, I., Koivisto, E., Riekkinen, P.J., 1991. The effects of guanfacine, alpha-2 agonist, on the performance of young and aged rats in spatial navigation task. Behav. Neural Biol. 56, 101–107. https://doi.org/10.1016/0163-1047(91)90327-m
Skvortsova, V., Palminteri, S., Pessiglione, M., 2014. Learning To Minimize Efforts versus Maximizing Rewards: Computational Principles and Neural Correlates. J. Neurosci. 34, 15621–15630. https://doi.org/10.1523/JNEUROSCI.1350-14.2014
Sloman, S.A., 1996. The empirical case for two systems of reasoning. Psychol. Bull. 119, 3–22. https://doi.org/10.1037/0033-2909.119.1.3
Smit, P.J., 1978. Sports medicine and “overtraining.” South Afr. Med. J. Suid-Afr. Tydskr. Vir Geneeskd. 54, 4.
Smith, C.C., Greene, R.W., 2012. CNS Dopamine Transmission Mediated by Noradrenergic Innervation. J. Neurosci. 32, 6072–6080. https://doi.org/10.1523/JNEUROSCI.6486-11.2012
Smith, S.A., Querry, R.G., Fadel, P.J., Gallagher, K.M., Strømstad, M., Ide, K., Raven, P.B., Secher, N.H., 2003. Partial blockade of skeletal muscle somatosensory afferents attenuates baroreflex resetting during exercise in humans. J. Physiol. 551, 1013–1021. https://doi.org/10.1113/jphysiol.2003.044925
Smith, V.L., Walker, J.M., 1993. Monetary Rewards and Decision Cost in Experimental Economics. Econ. Inq. 31, 245–261. https://doi.org/10.1111/j.1465-7295.1993.tb00881.x
Sofuoglu, M., Herman, A.I., Robinson, C., Waters, A.J., 2014. Chapter Twelve - Cognitive Effects of Nicotine, in: Madras, B., Kuhar, M. (Eds.), The Effects of Drug Abuse on the Human Nervous System. Academic Press, Boston, pp. 367–385. https://doi.org/10.1016/B978-0-12-418679-8.00012-5
Soltani, M., Knight, R.T., 2000. Neural origins of the P300. Crit. Rev. Neurobiol. 14, 199–224.
Sommer, M.A., 2003. The role of the thalamus in motor control. Curr. Opin. Neurobiol. 13, 663–670. https://doi.org/10.1016/j.conb.2003.10.014
Sonneborn, A., Greene, R.W., 2021. Norepinephrine transporter antagonism prevents dopamine-dependent synaptic plasticity in the mouse dorsal hippocampus.
References
297
Neurosci. Lett. 740, 135450. https://doi.org/10.1016/j.neulet.2020.135450
Sorg, O., Magistretti, P.J., 1992. Vasoactive intestinal peptide and noradrenaline exert long-term control on glycogen levels in astrocytes: blockade by protein synthesis inhibition. J. Neurosci. Off. J. Soc. Neurosci. 12, 4923–4931.
Sorg, O., Magistretti, P.J., 1991. Characterization of the glycogenolysis elicited by vasoactive intestinal peptide, noradrenaline and adenosine in primary cultures of mouse cerebral cortical astrocytes. Brain Res. 563, 227–233. https://doi.org/10.1016/0006-8993(91)91538-c
Stanovich, K.E., West, R.F., 2000. Individual differences in reasoning: Implications for the rationality debate? Behav. Brain Sci. 23, 645–665. https://doi.org/10.1017/S0140525X00003435
Stanton, P.K., Sarvey, J.M., 1985. Depletion of norepinephrine, but not serotonin, reduces long-term potentiation in the dentate gyrus of rat hippocampal slices. J. Neurosci. 5, 2169–2176. https://doi.org/10.1523/JNEUROSCI.05-08-02169.1985
Stefani, M.R., Groth, K., Moghaddam, B., 2003. Glutamate receptors in the rat medial prefrontal cortex regulate set-shifting ability. Behav. Neurosci. 117, 728–737. https://doi.org/10.1037/0735-7044.117.4.728
Stefani, M.R., Moghaddam, B., 2005. Systemic and Prefrontal Cortical NMDA Receptor Blockade Differentially Affect Discrimination Learning and Set-Shift Ability in Rats. Behav. Neurosci. 119, 420–428. https://doi.org/10.1037/0735-7044.119.2.420
Steindler, D.A., 1981. Locus coeruleus neurons have axons that branch to the forebrain and cerebellum. Brain Res. 223, 367–373. https://doi.org/10.1016/0006-8993(81)91149-5
Stephens, D.W., Krebs, J.R., 1986. Foraging theory, Monographs in behavior and ecology. Princeton University Press, Princeton, N.J.
Stevens, J.R., Rosati, A.G., Ross, K.R., Hauser, M.D., 2005. Will travel for food: spatial discounting in two new world monkeys. Curr. Biol. CB 15, 1855–1860. https://doi.org/10.1016/j.cub.2005.09.016
Stone, E.A., Zhang, Y., Hiller, J.M., Simon, E.J., Hillman, D.E., 1997. Activation of fos in mouse amygdala by local infusion of norepinephrine or atipamezole. Brain Res. 778, 1–5. https://doi.org/10.1016/S0006-8993(97)00667-7
Stone, E.A., Zhang, Y., John, S.M., Bing, G., 1991. C-Fos response to administration of catecholamines into brain by microdialysis. Neurosci. Lett. 133, 33–35. https://doi.org/10.1016/0304-3940(91)90050-4
Straube, T., Korz, V., Balschun, D., Frey, J.U., 2003. Requirement of beta-adrenergic receptor activation and protein synthesis for LTP-reinforcement by novelty in rat dentate gyrus. J. Physiol. 552, 953–960. https://doi.org/10.1113/jphysiol.2003.049452
Stroop, J.R., 1935. Studies of interference in serial verbal reactions. J. Exp. Psychol. 18, 643–662. https://doi.org/10.1037/h0054651
Sullivan, R., Wilson, D., Leon, M., 1989. Norepinephrine and learning-induced plasticity in infant rat olfactory system. J. Neurosci. 9, 3998–4006. https://doi.org/10.1523/JNEUROSCI.09-11-03998.1989
Sutin, E.L., Jacobowitz, D.M., 1991. Neurochemicals in the dorsal pontine tegmentum. Prog. Brain Res. 88, 3–14. https://doi.org/10.1016/s0079-6123(08)63796-6
Sved, A.F., Cano, G., Passerin, A.M., Rabin, B.S., 2002. The locus coeruleus, Barrington’s nucleus, and neural circuits of stress. Physiol. Behav. 77, 737–742. https://doi.org/10.1016/s0031-9384(02)00927-7
Swanson, L.W., 1976. The locus coeruleus: a cytoarchitectonic, Golgi and immunohistochemical study in the albino rat. Brain Res. 110, 39–56. https://doi.org/10.1016/0006-8993(76)90207-9
Swanson, R.A., 1992. Physiologic coupling of glial glycogen metabolism to neuronal activity in brain. Can. J. Physiol. Pharmacol. 70 Suppl, S138-144. https://doi.org/10.1139/y92-255
Swanson, R.A., Morton, M.M., Sagar, S.M., Sharp, F.R., 1992. Sensory stimulation induces local cerebral glycogenolysis: demonstration by
References
298
autoradiography. Neuroscience 51, 451–461. https://doi.org/10.1016/0306-4522(92)90329-z
Swick, D., Pineda, J.A., Schacher, S., Foote, S.L., 1994. Locus coeruleus neuronal activity in awake monkeys: relationship to auditory P300-like potentials and spontaneous EEG. Exp. Brain Res. 101, 86–92. https://doi.org/10.1007/BF00243219
Szabadi, E., 2013. Functional neuroanatomy of the central noradrenergic system. J. Psychopharmacol. Oxf. Engl. 27, 659–693. https://doi.org/10.1177/0269881113490326
Tait, D.S., Brown, V.J., Farovik, A., Theobald, D.E., Dalley, J.W., Robbins, T.W., 2007. Lesions of the dorsal noradrenergic bundle impair attentional set-shifting in the rat. Eur. J. Neurosci. 25, 3719–3724. https://doi.org/10.1111/j.1460-9568.2007.05612.x
Takeuchi, T., Duszkiewicz, A.J., Sonneborn, A., Spooner, P.A., Yamasaki, M., Watanabe, M., Smith, C.C., Fernández, G., Deisseroth, K., Greene, R.W., Morris, R.G.M., 2016. Locus coeruleus and dopaminergic consolidation of everyday memory. Nature 537, 357–362. https://doi.org/10.1038/nature19325
Talmi, D., Dayan, P., Kiebel, S.J., Frith, C.D., Dolan, R.J., 2009. How Humans Integrate the Prospects of Pain and Reward during Choice. J. Neurosci. 29, 14617–14626.
Tanila, H., Rämä, P., Carlson, S., 1996. The effects of prefrontal intracortical microinjections of an alpha-2 agonist, alpha-2 antagonist and lidocaine on the delayed alternation performance of aged rats. Brain Res. Bull. 40, 117–119. https://doi.org/10.1016/0361-9230(96)00026-3
Tattersall, A.J., Hockey, G.R.J., 1995. Level of operator control and changes in heart rate variability during simulated flight maintenance. Hum. Factors 37, 682–698. https://doi.org/10.1518/001872095778995517
Taylor, S.F., Welsh, R.C., Wager, T.D., Phan, K.L., Fitzgerald, K.D., Gehring, W.J., 2004. A functional neuroimaging study of motivation and executive function. NeuroImage 21, 1045–1054. https://doi.org/10.1016/j.neuroimage.2003.10.032
Tervo, D.G.R., Proskurin, M., Manakov, M., Kabra, M., Vollmer, A., Branson, K., Karpova, A.Y., 2014.
Behavioral Variability through Stochastic Choice and Its Gating by Anterior Cingulate Cortex. Cell 159, 21–32. https://doi.org/10.1016/j.cell.2014.08.037
Tibana, R., Manuel Frade de Sousa, N., Prestes, J., da Cunha Nascimento, D., Ernesto, C., Falk Neto, J., Kennedy, M., Azevedo Voltarelli, F., 2019. Is Perceived Exertion a Useful Indicator of the Metabolic and Cardiovascular Responses to a Metabolic Conditioning Session of Functional Fitness? Sports 7, 161. https://doi.org/10.3390/sports7070161
Tiesinga, P., Fellous, J.-M., Sejnowski, T.J., 2008. Regulation of spike timing in visual cortical circuits. Nat. Rev. Neurosci. 9, 97–107. https://doi.org/10.1038/nrn2315
Tillman, C.M., Wiens, S., 2011. Behavioral and ERP indices of response conflict in Stroop and flanker tasks. Psychophysiology 48, 1405–1411. https://doi.org/10.1111/j.1469-8986.2011.01203.x
Timberlake, W., 1994. Behavior systems, associationism, and Pavlovian conditioning. Psychon. Bull. Rev. 1, 405–420. https://doi.org/10.3758/BF03210945
Todd, M., Niv, Y., Cohen, J., 2008. Learning to Use Working Memory in Partially Observable Environments through Dopaminergic Reinforcement. Presented at the Advances in Neural Information Processing Systems, pp. 1689–1696.
Todd, P.M., Gigerenzer, G. (Eds.), 2012. Ecological rationality: intelligence in the world, Evolution and cognition series. Oxford University Press, Oxford ; New York.
Totah, N.K., Neves, R.M., Panzeri, S., Logothetis, N.K., Eschenko, O., 2018. The Locus Coeruleus Is a Complex and Differentiated Neuromodulatory System. Neuron 99, 1055-1068.e6. https://doi.org/10.1016/j.neuron.2018.07.037
Toyoda, H., Li, X.-Y., Wu, L.-J., Zhao, M.-G., Descalzi, G., Chen, T., Koga, K., Zhuo, M., 2011. Interplay of Amygdala and Cingulate Plasticity in Emotional Fear. Neural Plast. 2011, e813749. https://doi.org/10.1155/2011/813749
Travagli, R.A., Dunwiddie, T.V., Williams, J.T., 1995. Opioid Inhibition in Locus Coeruleus. J.
References
299
Neurophysiol. 74, 519–528. https://doi.org/10.1152/jn.1995.74.2.519
Treadway, M.T., Buckholtz, J.W., Schwartzman, A.N., Lambert, W.E., Zald, D.H., 2009. Worth the ‘EEfRT’? The Effort Expenditure for Rewards Task as an Objective Measure of Motivation and Anhedonia. PLOS ONE 4, e6598. https://doi.org/10.1371/journal.pone.0006598
Tsuda, K., Yokoo, H., Goldstein, M., 1989. Neuropeptide Y and galanin in norepinephrine release in hypothalamic slices. Hypertens. Dallas Tex 1979 14, 81–86. https://doi.org/10.1161/01.hyp.14.1.81
Tsukahara, J.S., Harrison, T.L., Engle, R.W., 2016. The relationship between baseline pupil size and intelligence. Cognit. Psychol. 91, 109–123. https://doi.org/10.1016/j.cogpsych.2016.10.001
Tsunematsu, S., 2001. Effort- and time-cost effects on demand curves for food by pigeons under short session closed economies. Behav. Processes 53, 47–56. https://doi.org/10.1016/S0376-6357(00)00147-9
Turetsky, B.I., Fein, G., 2002. α2-Noradrenergic effects on ERP and behavioral indices of auditory information processing. Psychophysiology 39, 147–157. https://doi.org/10.1111/1469-8986.3920147
Turner, J.R., Carroll, D., 1985. Heart Rate and Oxygen Consumption during Mental Arithmetic, a Video Game, and Graded Exercise: Further Evidence of Metabolically-Exaggerated Cardiac Adjustments? Psychophysiology 22, 261–267. https://doi.org/10.1111/j.1469-8986.1985.tb01597.x
Turner, J.R., Carroll, D., Courtney, H., 1983. Cardiac and Metabolic Responses to “Space Invaders”: An Instance of Metabolically-Exaggerated Cardiac Adjustment? Psychophysiology 20, 544–549. https://doi.org/10.1111/j.1469-8986.1983.tb03010.x
Tversky, A., Kahneman, D., 1974. Judgment under Uncertainty: Heuristics and Biases. Science 185, 1124–1131. https://doi.org/10.1126/science.185.4157.1124
Tzavara, E.T., Bymaster, F.P., Overshiner, C.D., Davis, R.J., Perry, K.W., Wolff, M., McKinzie, D.L.,
Witkin, J.M., Nomikos, G.G., 2006. Procholinergic and memory enhancing properties of the selective norepinephrine uptake inhibitor atomoxetine. Mol. Psychiatry 11, 187–195. https://doi.org/10.1038/sj.mp.4001763
Uematsu, A., Tan, B.Z., Johansen, J.P., 2015. Projection specificity in heterogeneous locus coeruleus cell populations: implications for learning and memory. Learn. Mem. 22, 444–451. https://doi.org/10.1101/lm.037283.114
Uematsu, A., Tan, B.Z., Ycu, E.A., Cuevas, J.S., Koivumaa, J., Junyent, F., Kremer, E.J., Witten, I.B., Deisseroth, K., Johansen, J.P., 2017. Modular organization of the brainstem noradrenaline system coordinates opposing learning states. Nat. Neurosci. 20, 1602–1611. https://doi.org/10.1038/nn.4642
Uhr, M.L., 1979. The influence of an enzyme on the direction of a reaction. Biochem. Educ. 7, 15–16. https://doi.org/10.1016/0307-4412(79)90014-1
Ullsperger, M., Cramon, D.Y. von, 2003. Error Monitoring Using External Feedback: Specific Roles of the Habenular Complex, the Reward System, and the Cingulate Motor Area Revealed by Functional Magnetic Resonance Imaging. J. Neurosci. 23, 4308–4314. https://doi.org/10.1523/JNEUROSCI.23-10-04308.2003
Umemoto, A., Holroyd, C.B., 2015. Task-specific effects of reward on task switching. Psychol. Res. 79, 698–707. https://doi.org/10.1007/s00426-014-0595-z
U’Prichard, D.C., Bechtel, W.D., Rouot, B.M., Snyder, S.H., 1979. Multiple apparent alpha-noradrenergic receptor binding sites in rat brain: effect of 6-hydroxydopamine. Mol. Pharmacol. 16, 47–60.
Usher, M., Cohen, J.D., Servan-Schreiber, D., Rajkowski, J., Aston-Jones, G., 1999. The Role of Locus Coeruleus in the Regulation of Cognitive Performance. Science 283, 549–554. https://doi.org/10.1126/science.283.5401.549
Usher, M., Davelaar, E.J., 2002. Neuromodulation of decision and response selection. Neural Netw. 15, 635–645. https://doi.org/10.1016/S0893-6080(02)00054-0
References
300
Valentin, V.V., Dickinson, A., O’Doherty, J.P., 2007. Determining the Neural Substrates of Goal-Directed Learning in the Human Brain. J. Neurosci. 27, 4019–4026.
Valentino, R.J., Foote, S.L., 1988. Corticotropin-releasing hormone increases tonic but not sensory-evoked activity of noradrenergic locus coeruleus neurons in unanesthetized rats. J. Neurosci. Off. J. Soc. Neurosci. 8, 1016–1025.
Valentino, R.J., Page, M.E., Curtis, A.L., 1991. Activation of noradrenergic locus coeruleus neurons by hemodynamic stress is due to local release of corticotropin-releasing factor. Brain Res. 555, 25–34. https://doi.org/10.1016/0006-8993(91)90855-p
Valentino, R.J., Van Bockstaele, E., 2008. Convergent Regulation of Locus Coeruleus Activity as an adaptive Response to Stress. Eur. J. Pharmacol. 583, 194–203. https://doi.org/10.1016/j.ejphar.2007.11.062
Van Bockstaele, E.J., Aston-Jones, G., 1995. Integration in the Ventral Medulla and Coordination of Sympathetic, Pain and Arousal Functions. Clin. Exp. Hypertens. 17, 153–165. https://doi.org/10.3109/10641969509087062
Van Bockstaele, E.J., Bajic, D., Proudfit, H., Valentino, R.J., 2001. Topographic architecture of stress-related pathways targeting the noradrenergic locus coeruleus. Physiol. Behav. 73, 273–283. https://doi.org/10.1016/s0031-9384(01)00448-6
Van Cutsem, J., Marcora, S., De Pauw, K., Bailey, S., Meeusen, R., Roelands, B., 2017. The Effects of Mental Fatigue on Physical Performance: A Systematic Review. Sports Med. 47, 1569–1588. https://doi.org/10.1007/s40279-016-0672-0
van der Wel, P., van Steenbergen, H., 2018. Pupil dilation as an index of effort in cognitive control tasks: A review. Psychon. Bull. Rev. 25, 2005–2015. https://doi.org/10.3758/s13423-018-1432-y
van Marle, H.J.F., Hermans, E.J., Qin, S., Fernández, G., 2010. Enhanced resting-state connectivity of amygdala in the immediate aftermath of acute psychological stress. NeuroImage 53, 348–354. https://doi.org/10.1016/j.neuroimage.2010.05.070
van Veen, V., Carter, C.S., 2002. The Timing of Action-Monitoring Processes in the Anterior Cingulate Cortex. J. Cogn. Neurosci. 14, 593–602. https://doi.org/10.1162/08989290260045837
van Veen, V., Cohen, J.D., Botvinick, M.M., Stenger, V.A., Carter, C.S., 2001. Anterior Cingulate Cortex, Conflict Monitoring, and Levels of Processing. NeuroImage 14, 1302–1308. https://doi.org/10.1006/nimg.2001.0923
Vankov, A., Hervé-Minvielle, A., Sara, S.J., 1995. Response to novelty and its rapid habituation in locus coeruleus neurons of the freely exploring rat. Eur. J. Neurosci. 7, 1180–1187. https://doi.org/10.1111/j.1460-9568.1995.tb01108.x
Varazzani, C., San-Galli, A., Gilardeau, S., Bouret, S., 2015. Noradrenaline and Dopamine Neurons in the Reward/Effort Trade-Off: A Direct Electrophysiological Comparison in Behaving Monkeys. J. Neurosci. 35, 7866–7877. https://doi.org/10.1523/JNEUROSCI.0454-15.2015
Vazey, E.M., Moorman, D.E., Aston-Jones, G., 2018. Phasic locus coeruleus activity regulates cortical encoding of salience information. Proc. Natl. Acad. Sci. U. S. A. 115, E9439–E9448. https://doi.org/10.1073/pnas.1803716115
Ventura, R., Latagliata, E.C., Morrone, C., La Mela, I., Puglisi-Allegra, S., 2008. Prefrontal Norepinephrine Determines Attribution of “High” Motivational Salience. PLoS ONE 3, e3044. https://doi.org/10.1371/journal.pone.0003044
Ventura, R., Morrone, C., Puglisi-Allegra, S., 2007. Prefrontal/accumbal catecholamine system determines motivational salience attribution to both reward- and aversion-related stimuli. Proc. Natl. Acad. Sci. 104, 5181–5186. https://doi.org/10.1073/pnas.0610178104
Verguts, T., Vassena, E., Silvetti, M., 2015. Adaptive effort investment in cognitive and physical tasks: a neurocomputational model. Front. Behav. Neurosci. 9. https://doi.org/10.3389/fnbeh.2015.00057
Verleger, R., Heide, W., Butt, C., Kömpf, D., 1994. Reduction of P3b in patients with temporo-parietal lesions. Brain Res. Cogn. Brain Res. 2, 103–116. https://doi.org/10.1016/0926-6410(94)90007-8
References
301
Vissing, J., Andersen, M., Diemer, N.H., 1996. Exercise-induced changes in local cerebral glucose utilization in the rat. J. Cereb. Blood Flow Metab. Off. J. Int. Soc. Cereb. Blood Flow Metab. 16, 729–736. https://doi.org/10.1097/00004647-199607000-00025
Von Neumann, J., Morgenstern, O., 2007. Theory of games and economic behavior, 60th anniversary ed. ed, Princeton classic editions. Princeton University Press, Princeton, N.J. ; Woodstock.
Von Neumann, J., Morgenstern, O., 1947. Theory of games and economic behavior, 2nd rev. ed, Theory of games and economic behavior, 2nd rev. ed. Princeton University Press, Princeton, NJ, US.
Wagatsuma, A., Okuyama, T., Sun, C., Smith, L.M., Abe, K., Tonegawa, S., 2018. Locus coeruleus input to hippocampal CA3 drives single-trial learning of a novel context. Proc. Natl. Acad. Sci. U. S. A. 115, E310–E316. https://doi.org/10.1073/pnas.1714082115
Wahn, B., Ferris, D.P., Hairston, W.D., König, P., 2016. Pupil Sizes Scale with Attentional Load and Task Experience in a Multiple Object Tracking Task. PloS One 11, e0168087. https://doi.org/10.1371/journal.pone.0168087
Walling, S.G., Harley, C.W., 2004. Locus ceruleus activation initiates delayed synaptic potentiation of perforant path input to the dentate gyrus in awake rats: a novel beta-adrenergic- and protein synthesis-dependent mammalian plasticity mechanism. J. Neurosci. Off. J. Soc. Neurosci. 24, 598–604. https://doi.org/10.1523/JNEUROSCI.4426-03.2004
Walls, A.B., Heimbürger, C.M., Bouman, S.D., Schousboe, A., Waagepetersen, H.S., 2009. Robust glycogen shunt activity in astrocytes: Effects of glutamatergic and adrenergic agents. Neuroscience 158, 284–292. https://doi.org/10.1016/j.neuroscience.2008.09.058
Walton, M.E., Bannerman, D.M., Alterescu, K., Rushworth, M.F.S., 2003. Functional Specialization within Medial Frontal Cortex of the Anterior Cingulate for Evaluating Effort-Related Decisions. J. Neurosci. 23, 6475–6479.
https://doi.org/10.1523/JNEUROSCI.23-16-06475.2003
Walton, M.E., Bannerman, D.M., Rushworth, M.F.S., 2002. The role of rat medial frontal cortex in effort-based decision making. J. Neurosci. Off. J. Soc. Neurosci. 22, 10996–11003.
Walton, M.E., Croxson, P.L., Rushworth, M.F.S., Bannerman, D.M., 2005. The Mesocortical Dopamine Projection to Anterior Cingulate Cortex Plays No Role in Guiding Effort-Related Decisions. Behav. Neurosci. 119, 323–328. https://doi.org/10.1037/0735-7044.119.1.323
Walton, M.E., Devlin, J.T., Rushworth, M.F.S., 2004. Interactions between decision making and performance monitoring within prefrontal cortex. Nat. Neurosci. 7, 1259–1265. https://doi.org/10.1038/nn1339
Walton, M.E., Groves, J., Jennings, K.A., Croxson, P.L., Sharp, T., Rushworth, M.F.S., Bannerman, D.M., 2009. Comparing the role of the anterior cingulate cortex and 6-hydroxydopamine nucleus accumbens lesions on operant effort-based decision making. Eur. J. Neurosci. 29, 1678–1691. https://doi.org/10.1111/j.1460-9568.2009.06726.x
Walton, M.E., Kennerley, S.W., Bannerman, D.M., Phillips, P.E.M., Rushworth, M.F.S., 2006. Weighing up the benefits of work: behavioral and neural analyses of effort-related decision making. Neural Netw. Off. J. Int. Neural Netw. Soc. 19, 1302–1314. https://doi.org/10.1016/j.neunet.2006.03.005
Wang, M., Ramos, B.P., Paspalas, C.D., Shu, Y., Simen, A., Duque, A., Vijayraghavan, S., Brennan, A., Dudley, A., Nou, E., Mazer, J.A., McCormick, D.A., Arnsten, A.F.T., 2007. α2A-Adrenoceptors Strengthen Working Memory Networks by Inhibiting cAMP-HCN Channel Signaling in Prefrontal Cortex. Cell 129, 397–410. https://doi.org/10.1016/j.cell.2007.03.015
Wang, X.-J., 2002. Probabilistic Decision Making by Slow Reverberation in Cortical Circuits. Neuron 36, 955–968. https://doi.org/10.1016/S0896-6273(02)01092-9
Watanabe, H., Passonneau, J.V., 1973. Factors affecting the turnover of cerebral glycogen and limit dextrin in vivo. J. Neurochem. 20, 1543–
References
302
1554. https://doi.org/10.1111/j.1471-4159.1973.tb00272.x
Waterhouse, B.D., Moises, H.C., Woodward, D.J., 1998. Phasic activation of the locus coeruleus enhances responses of primary sensory cortical neurons to peripheral receptive field stimulation. Brain Res. 790, 33–44. https://doi.org/10.1016/s0006-8993(98)00117-6
Waterhouse, B.D., Moises, H.C., Woodward, D.J., 1980. Noradrenergic modulation of somatosensory cortical neuronal responses to lontophoretically applied putative neurotransmitters. Exp. Neurol. 69, 30–49. https://doi.org/10.1016/0014-4886(80)90141-7
Waterhouse, B.D., Mouradian, R., Sessler, F.M., Lin, R.C.S., 2000. Differential modulatory effects of norepinephrine on synaptically driven responses of layer V barrel field cortical neurons. Brain Res. 868, 39–47. https://doi.org/10.1016/S0006-8993(00)02261-7
Waterhouse, B.D., Sessler, F.M., Jung-Tung, C., Woodward, D.J., Azizi, S.A., Moises, H.C., 1988. New evidence for a gating action of norepinephrine in central neuronal circuits of mammalian brain. Brain Res. Bull. 21, 425–432. https://doi.org/10.1016/0361-9230(88)90154-2
Waterhouse, B.D., Woodward, D.J., 1980. Interaction of norepinephrine with cerebrocortical activity evoked by stimulation of
somatosensory afferent pathways in the rat☆. Exp. Neurol. 67, 11–34. https://doi.org/10.1016/0014-4886(80)90159-4
Wei, H., Zhou, R., 2020. High working memory load impairs selective attention: EEG signatures. Psychophysiology 57, e13643. https://doi.org/10.1111/psyp.13643
Westbrook, A., Braver, T.S., 2015. Cognitive effort: A neuroeconomic approach. Cogn. Affect. Behav. Neurosci. 15, 395–415. https://doi.org/10.3758/s13415-015-0334-y
Westbrook, A., Kester, D., Braver, T.S., 2013. What Is the Subjective Cost of Cognitive Effort? Load, Trait, and Aging Effects Revealed by Economic Preference. PLOS ONE 8, e68210. https://doi.org/10.1371/journal.pone.0068210
Westbrook, J.A., Braver, T.S., 2013. The economics of cognitive effort. Behav. Brain Sci.
36, 704–705; discussion 707-726. https://doi.org/10.1017/S0140525X13001179
White, O., Davare, M., Andres, M., Olivier, E., 2013. The role of left supplementary motor area in grip force scaling. PloS One 8, e83812. https://doi.org/10.1371/journal.pone.0083812
Whittington, M.A., Cunningham, M.O., LeBeau, F.E.N., Racca, C., Traub, R.D., 2011. Multiple origins of the cortical gamma rhythm. Dev. Neurobiol. 71, 92–106. https://doi.org/10.1002/dneu.20814
Wickelgren, W.A., 1977. Speed-accuracy tradeoff and information processing dynamics. Acta Psychol. (Amst.) 41, 67–85. https://doi.org/10.1016/0001-6918(77)90012-9
Widge, A.S., Zorowitz, S., Basu, I., Paulk, A.C., Cash, S.S., Eskandar, E.N., Deckersbach, T., Miller, E.K., Dougherty, D.D., 2019. Deep brain stimulation of the internal capsule enhances human cognitive control and prefrontal cortex function. Nat. Commun. 10, 1536. https://doi.org/10.1038/s41467-019-09557-4
Wiecki, T.V., Frank, M.J., 2013. A computational model of inhibitory control in frontal cortex and basal ganglia. Psychol. Rev. 120, 329–355. https://doi.org/10.1037/a0031542
Williams, J.T., North, R.A., Shefner, S.A., Nishi, S., Egan, T.M., 1984. Membrane properties of rat locus coeruleus neurones. Neuroscience 13, 137–156. https://doi.org/10.1016/0306-4522(84)90265-3
Wilson, G.F., 1992. Applied use of cardiac and respiration measures: Practical considerations and precautions. Biol. Psychol. 34, 163–178. https://doi.org/10.1016/0301-0511(92)90014-L
Witte, E.A., Marrocco, R.T., 1997. Alteration of brain noradrenergic activity in rhesus monkeys affects the alerting component of covert orienting. Psychopharmacology (Berl.) 132, 315–323. https://doi.org/10.1007/s002130050351
Womelsdorf, T., Schoffelen, J.-M., Oostenveld, R., Singer, W., Desimone, R., Engel, A.K., Fries, P., 2007. Modulation of neuronal interactions through neuronal synchronization. Science 316, 1609–1612. https://doi.org/10.1126/science.1139597
References
303
Worthy, D.A., Markman, A.B., Maddox, W.T., 2009. What is pressure? Evidence for social pressure as a type of regulatory focus. Psychon. Bull. Rev. 16, 344–349. https://doi.org/10.3758/PBR.16.2.344
Wright, P., 1975. Consumer Choice Strategies: Simplifying Vs. Optimizing. J. Mark. Res. 12, 60–67. https://doi.org/10.1177/002224377501200109
Wright, T.J., Boot, W.R., Morgan, C.S., 2013. Pupillary response predicts multiple object tracking load, error rate, and conscientiousness, but not inattentional blindness. Acta Psychol. (Amst.) 144, 6–11. https://doi.org/10.1016/j.actpsy.2013.04.018
Xu, Z.Q., Shi, T.J., Hökfelt, T., 1998. Galanin/GMAP- and NPY-like immunoreactivities in locus coeruleus and noradrenergic nerve terminals in the hippocampal formation and cortex with notes on the galanin-R1 and -R2 receptors. J. Comp. Neurol. 392, 227–251. https://doi.org/10.1002/(sici)1096-9861(19980309)392:2<227::aid-cne6>3.0.co;2-4
Yerkes, R.M., Dodson, J.D., 1908. The Relation of Strength of Stimulus to Rapidity of Habit Formation. J. Comp. Neurol. Psychol. 18, 459–482. https://doi.org/10.1002/cne.920180503
Yingling, C.D., Hosobuchi, Y., 1984. A subcortical correlate of P300 in man. Electroencephalogr. Clin. Neurophysiol. Evoked Potentials 59, 72–76. https://doi.org/10.1016/0168-5597(84)90022-4
Yoshida, W., Ishii, S., 2006. Resolution of uncertainty in prefrontal cortex. Neuron 50, 781–789. https://doi.org/10.1016/j.neuron.2006.05.006
Young, W.S., Kuhar, M.J., 1980. Noradrenergic alpha 1 and alpha 2 receptors: light microscopic autoradiographic localization. Proc. Natl. Acad. Sci. U. S. A. 77, 1696–1700. https://doi.org/10.1073/pnas.77.3.1696
Yu, A.J., Dayan, P., 2005. Uncertainty, Neuromodulation, and Attention. Neuron 46, 681–692. https://doi.org/10.1016/j.neuron.2005.04.026
Yu, A.J., Dayan, P., Cohen, J.D., 2009. Dynamics of attentional selection under conflict: Toward a rational Bayesian account. J. Exp. Psychol. Hum.
Percept. Perform. 35, 700–717. https://doi.org/10.1037/a0013553
Yuen, E.Y., Liu, W., Karatsoreos, I.N., Ren, Y., Feng, J., McEwen, B.S., Yan, Z., 2011. Mechanisms for acute stress-induced enhancement of glutamatergic transmission and working memory. Mol. Psychiatry 16, 156–170. https://doi.org/10.1038/mp.2010.50
Yuen, E.Y., Wei, J., Liu, W., Zhong, P., Li, X., Yan, Z., 2012. Repeated Stress Causes Cognitive Impairment by Suppressing Glutamate Receptor Expression and Function in Prefrontal Cortex. Neuron 73, 962–977. https://doi.org/10.1016/j.neuron.2011.12.033
Zénon, A., Sidibé, M., Olivier, E., 2015. Disrupting the Supplementary Motor Area Makes Physical Effort Appear Less Effortful. J. Neurosci. 35, 8737–8744. https://doi.org/10.1523/JNEUROSCI.3789-14.2015
Zénon, A., Sidibé, M., Olivier, E., 2014. Pupil size variations correlate with physical effort perception. Front. Behav. Neurosci. 8, 286. https://doi.org/10.3389/fnbeh.2014.00286
Zerbi, V., Floriou-Servou, A., Markicevic, M., Vermeiren, Y., Sturman, O., Privitera, M., von Ziegler, L., Ferrari, K.D., Weber, B., De Deyn, P.P., Wenderoth, N., Bohacek, J., 2019. Rapid Reconfiguration of the Functional Connectome after Chemogenetic Locus Coeruleus Activation. Neuron 103, 702-718.e5. https://doi.org/10.1016/j.neuron.2019.05.034
Zhao, W., Martin, A.D., Davenport, P.W., 2003. Magnitude estimation of inspiratory resistive loads by double-lung transplant recipients. J. Appl. Physiol. 94, 576–582. https://doi.org/10.1152/japplphysiol.00564.2002
Zhu, L., Mathewson, K.E., Hsu, M., 2012. Dissociable neural representations of reinforcement and belief prediction errors underlie strategic learning. Proc. Natl. Acad. Sci. U. S. A. 109, 1419–1424. https://doi.org/10.1073/pnas.1116783109
Zink, N., Bensmann, W., Arning, L., Stock, A.-K., Beste, C., 2019. CHRM2 Genotype Affects Inhibitory Control Mechanisms During Cognitive Flexibility. Mol. Neurobiol. 56, 6134–6141. https://doi.org/10.1007/s12035-019-1521-6
304
Zipf, G.K., 1949. Human behavior and the principle of least effort, Human behavior and the principle of least effort. Addison-Wesley Press, Oxford, England.
Zuend, M., Saab, A.S., Wyss, M.T., Ferrari, K.D., Hösli, L., Looser, Z.J., Stobart, J.L., Duran, J., Guinovart, J.J., Barros, L.F., Weber, B., 2020. Arousal-induced cortical activity triggers lactate
release from astrocytes. Nat. Metab. 2, 179–191. https://doi.org/10.1038/s42255-020-0170-4
Zwosta, K., Ruge, H., Wolfensteller, U., 2015. Neural mechanisms of goal-directed behavior: outcome-based response selection is associated with increased functional coupling of the angular gyrus. Front. Hum. Neurosci. 9. https://doi.org/10.3389/fnhum.2015.00180
List of figures
305
List of figures
Figure 1 : Physical effort carries disutility and increased effort requirements need to be
associated with increased rewards to be chosen (adapted from Walton et al., 2006) ............. 8
Figure 2 : Acceptance rate decreased when physical effort increased in a force exertion task
consisting in pressing a grip at four different force levels (adapted from Zénon et al., 2014) 10
Figure 3 : During force exertion, after an initial dip, pupil diameter increased and was positively
correlated with the exerted force (figure adapted from Zénon et al., 2014). ......................... 11
Figure 4: Positive relationship between RPEs and (A) oxygen uptake (percentage of maximum
uptake, VO2max) and (B) heart rate (percentage of maximum heart rate, HRmax) (adapted from
Andrade et al., 2020). ............................................................................................................... 12
Figure 5: Tasks used to test the different components of cognitive control ........................... 14
Figure 6: Pupil diameter increases during decisions, especially when subjects will respond
against their bias. ..................................................................................................................... 17
Figure 7: Behavioral demonstration of the aversiveness of mental effort (figure adapted from
Kool and Botvinick, 2018) ......................................................................................................... 22
Figure 8 : Results of the modified Stroop task used by Padmala and Pessoa (2011). RTs were
shorter, accuracy was higher and difficulty had lesser effects on RT and accuracy when
subjects were rewarded. .......................................................................................................... 23
Figure 9: Cognitive effort titration task (adapted from Westbrook et al., 2013). ................... 24
Figure 10 : The VOC model and the EVC model ....................................................................... 26
Figure 11: Representation of the determination of the optimal control intensity signal in a
certain state (payoffs and costs) and in case of increases in payoff or in difficulty (adapted from
Shenhav et al., 2013). ............................................................................................................... 28
Figure 12: Labor-leisure theory, in economics and applied to cognitive control (figure adapted
from Kool and Botvinick, 2014). ............................................................................................... 29
Figure 13 : Schematic representation of the different constraints on control exertion that
could account for limited control capacities and control aversiveness (adapted from Shenhav
et al., 2017). .............................................................................................................................. 32
Figure 14 : Ego-depletion after a controlled attention task, in a Stroop task with or without
glucose consumption (adapted from Gailliot et al., 2007). ..................................................... 34
Figure 15 : Glycogen as a buffer for brain metabolic resources (adapted from Christie and
Schrater, 2015) ......................................................................................................................... 35
List of figures
306
Figure 16: Influence of value difference (VD) and trial overall value (OV) or avoidance ratings
and on RTs in a value-based decision-making task .................................................................. 41
Figure 17 : The multiple-object tracking task (adapted from Alnæs et al., 2014). .................. 45
Figure 18: ERP recorded in EEG. ERPs can be decomposed into several subcomponents: N1
(positive), P2 (negative), N2 (positive) and P3 (negative) ....................................................... 47
Figure 19: Anatomy and histology of the LC in humans and monkeys .................................... 51
Figure 20: LC neurons respond to surprising stimuli with a characteristic activation-inhibition
response ................................................................................................................................... 52
Figure 21: ERPs evoked by LC optogenetic stimulation (adapted from Vazey et al., 2018). ... 58
Figure 22: Positive correlation between pupil diameter and LC activity ................................. 60
Figure 23 : Optogenetic inhibition of LC activity promotes sleep while stimulation promotes
wake (Adapted from Carter et al., 2010). ................................................................................ 62
Figure 24 : LC activity is modulated by stress, via the action of CRH. ..................................... 64
Figure 25: Optogenetic manipulations and Calcium imaging of LC activity demonstrate a causal
role of the LC/NA system in inhibitory control exertion (adapted from Bari et al., 2020) ...... 68
Figure 26: Pupil diameter increased more before exploration than exploitation, and individual
pupil diameter baseline was positively correlated with individual tendency to explore (adapted
from Jepma and Nieuwenhuis, 2011). ..................................................................................... 70
Figure 27: Increasing LC activity caused increased exploration and decision variability, while
decreasing LC activity decreased choice variability. ................................................................ 72
Figure 28: LC neurons encoded the exerted force during force exertion, and decreasing brain
NA decreased exerted force. .................................................................................................... 73
Figure 29: Effects of LC activity on action selection: increased LC activity, measured indirectly
through pupil dilation, was associated to decreased decision biases, while decreases in LC
activity by clonidine increased physical effort sensitivity. ....................................................... 76
Figure 30: LC neurons encoded engagement in the current trial at the time of cue onset and
in the next trial around fixation breaks i.e. disengagement from trials (adapted from Jahn et
al., 2020). .................................................................................................................................. 78
Figure 31: LC neurons differentiated between cues carrying novel information and cues
carrying no novel information. ................................................................................................. 80
Figure 32: LC neurons encoded some of the information provided by task cues on trial
contingencies, especially reward size. ..................................................................................... 81
List of figures
307
Figure 33: In forced-choice discrimination tasks, LC activity was better aligned on response
onset than on stimulus onset. LC neurons were only active if an action was executed,
regardless of reward expectation. ........................................................................................... 83
Figure 34: LC neurons encode task parameters just prior to action onset: reward negatively
and force positively. ................................................................................................................. 85
Figure 35: LC neurons respond to task-relevant cues. ............................................................. 91
Figure 36 : Effects of LC stimulation on activity of simultaneously recorded thalamic neurons
and somatosensory neurons. ................................................................................................... 93
Figure 37: In a spatial memory task, α2 agonist guanfacine enhanced while α2 antagonist
yohimbine decreased spatial-related responses of PFC neurons during memory retain period.
.................................................................................................................................................. 94
Figure 38: Yerkes-Dodson relationship between LC discharge and performance. .................. 96
Figure 39: The activation function, relating the input of a neuronal system (x-axis, Net input)
to the probability of firing in response to the input (y-axis, activation). ................................. 98
Figure 40: Network reset theory (adapted from Bouret and Sara, 2005). ............................ 102
Figure 41: Summary of the mean variations in rate for the 75 recorded neurons around cue
onset (A, B) and the outcome (C, D). ..................................................................................... 154
Figure 42: Summary of the modulation of rate by task parameters in large time intervals
around cue onset and the outcome ....................................................................................... 156
Figure 43: Summary of the dynamics of the modulation of rate by task parameters (Reward
and Delay) around cue onset (A,B,C), action onset (D, E, F) and the outcome (G, H, I). ....... 161
Figure 44: Summary of the correlations between baseline rate and the strength of the
response to task events and of modulation of rate by task parameters around task events.
................................................................................................................................................ 163
Figure 45: Two alternatives for the relationship between baseline rate and signal-to noise
ratio, represented using schematic representations of the mean rate of virtual example units
across time around an event of interest. ............................................................................... 165
Figure 46: Two-alternative choice task. ................................................................................. 175
Figure 47: Electrophysiological recordings: location in the brain and pharmacological
validation. ............................................................................................................................... 177
Figure 48: RT determination and decomposition: forces exerted in the two-alternative choice
task and eye-tracking procedure. .......................................................................................... 179
List of figures
308
Figure 49: Behavior associated to the middle press was compatible with a positive relationship
between RT and the cost of triggering the middle press. ...................................................... 184
Figure 50: Decisions to engage in the choice in the two-alternative choice task. ................. 186
Figure 51: Action-triggering time increased with difficulty to engage in the choice ............. 188
Figure 52: The subjective difficulty to engage in the choice scaled with the subjective difficulty
to engage in the middle press. ............................................................................................... 191
Figure 53: Binary choices between offers, and action-selection times. ................................ 194
Figure 54: Relationship between action-triggering time and choice accuracy: tenuous evidence
for a speed-accuracy tradeoff, and evidence for a “speed-cognitive control production”
tradeoff. .................................................................................................................................. 200
Figure 55: Negative relationship between maximum exerted force and engagement difficulty.
................................................................................................................................................ 203
Figure 56: Activity of LC neurons around middle press onset: LC neurons activated around the
middle press and encoded middle press RT before the middle press. .................................. 210
Figure 57: Increased LC activity before middle press onset was associated to decreased
difficulty to engage in the choice, as choice engagement increased and action-triggering time
decreased. .............................................................................................................................. 214
Figure 58: Pre-middle press spike count had a positive relation with middle press exerted
force........................................................................................................................................ 216
Figure 59: Rate of LC neurons around cues onset, and results of first and second order t-tests
assessing rate changes: after cues onset, LC activity increased. ........................................... 218
Figure 60: LC neurons encoded RT positively, and specifically its subcomponent action-
triggering time, over and above offer parameters, around cues onset. ............................... 220
Figure 61: Encoding of offer parameters (Forces and Rewards) in a late epoch ([200:500ms])
after cue onset: LC neurons only encoded Rewards. ............................................................. 223
Figure 62: LC activity around cues onset had a positive influence on the probability to choose
high-Force options, on the exerted force, especially in high-chosen Force conditions, and a
positive effect on performance. ............................................................................................. 225
Figure 63: Rate of LC neurons around side action onset, and results of first and second order
t-tests assessing rate changes. ............................................................................................... 231
Figure 64: LC neurons encoded the difficulty of the action during its execution. ................. 233
309
Figure 65: Increased spike count during the previous side press was associated to increased
Force sensitivity of the next Force-based choice. .................................................................. 235
Figure 66: LC neurons encoded performance of previous side press, but not engagement in
previous side press during the ITI ([-450;-150ms] before middle red dot onset). ................ 237
List of tables
Table 1: Latencies of the firing rate changes and task parameter encoding at cue onset,
action onset and the outcome. The latencies indicated correspond to the median and
confidence interval of the position of the windows in which a change was detected. ........ 159
Table 2: Results for the Pearson’s correlations of baseline rate with effect sizes
(maximum of the absolute values of coefficients from the spike count GLM) for reward, delay,
RT and willingness to work, at cue onset, action onset and the outcome. ........................... 166
List of abbreviations
310
List of abbreviations 4V: fourth ventricule
4x: trochlear decussation
ACC: anterior cingulate cortex
ADHD: attention-deficit hyperactivity disorder
AIC: Akaike Information Criteria
ANOVA: analysis of variance
AST: action-selection time
bc: brachium conjonctivium
BIC: Bayesian Information Criteria
BOLD: blood-oxygen-level dependent
cAMP: cyclic adenosine monophosphate
Cb: cerebellum
Cb2: cerebellum
CGPn: central gray of the rhombencephalon
ChR2: channelrhodopsin-2
CNO: clozapine-N-oxide
Corr: correct
CRH: corticotropin-release hormone
CS: conditioned stimulus
ctg: central tegmental tract
ctt: constant
D: delay
DA: dopamine
dACC: dorsal anterior cingulate cortex
DDM: drift-diffusion model
DREADD: Designer Receptor Exclusively Activated
by Designer Drugs
DβH: dopamine-β-hydroxylase
EEG: Electroencephalograms
eNpHR: halorhodopsin
ERP: event-related potential
EVC expected value of control
eYFP: yellow fluorescent protein
F: force
FA: false alarm
fMRI: functional magnetic resonance imaging
FR: number of lever presses
G-6-P: glucose-6-phosphate
GANE: glutamate amplification of noradrenaline
GLM: generalized linear model
HCN channel: hyperpolarization-activated cyclic
nucleotide gated cation channel
HF: high-force
HR: high-reward
HRmax: maximum heart rate
incorr: incorrect
ITI: inter-trial interval
LC: locus coeruleus
LR: low-reward
LTD: long-term depression
LTP: long-term potentiation
LV: lateral ventricule
mcp: middle cerebellar pedoncule
Me5: mesencephalic trigeminal nucleus
mlf: medial longitudinal fasciculus
Mo5N: rostral tip of the motor trigeminal nucleus
MPB: medial parabrachial nucleus
mPFC: medial prefrontal cortex
MRI: magnetic resonance imaging
mRNA: messenger ribonucleic acid
MVC: maximum voluntary contraction
n.s.: non significant
NA: noradrenaline
NAc: nucleus accumbens
NMDA: N-methyl-D-aspartate
NREM: non rapid eye movement
OV: overall value
PFC: prefrontal cortex
Pgi: nucleus Paragigantocellularis
Pn: pontine nuclei
PnO: pontine nuclei
prem: premature
PTG: protein targeting glycogen
PTSD: post traumatic stress disorder
R: reward
REM: rapid eye movement
RPE: rate of perceived exertion
RT: reaction time
scp: superior cerebellar pedoncule
SEM: standard error of the mean
SMA: supplementaru motor area
spk: spike
SSRT: stop signal reaction time
subCD: dorsal part of the subcoeruleus
subCV: ventral part pf the subcoeruleus
TPR: task-evoked pupillary responses
tVNS: transcutaneous vagus nerve stimulation
TW: time window
V: value
VD: value difference
ven: fourth ventricule
vmPFC: ventro-medial prefrontal cortex
VO2max: maximum oxygen uptake
VOC: value of computations
vsc: ventrat spinocerebellar tract
WTW: willingness to work
Résumé en français
311
Résumé en français
Allez-vous aller faire l’effort d’aller courir ce soir, ou resterez-vous dans votre canapé
devant un film? La prochaine fois que vous ferez vos courses, allez-vous faire l’effort de
calculer de tête le montant de tous vos articles ? Chaque jour, nous réalisons de nombreuses
décisions, et une grande partie d’entre elles repose sur un arbitrage entre un effort, physique
ou mental, et une récompense. L’effort est aversif, c’est-à-dire que les sujets agissent de telle
manière à éviter les actions qui impliqueraient de faire un effort. De plus, l’effort représente
les ressources qui sont mises en œuvre pour réaliser les actions, ce qui implique que faire un
effort est bénéfique pour la performance. Dans l’introduction du manuscrit, sont détaillées les
caractéristiques des deux types d’effort, physique et mental, ainsi que le lien entre ces deux
types d’effort, et dans un second temps, les éléments qui ont amené à la formulation de
l’hypothèse de l’implication du locus coeruleus (LC), noyau noradrénergique cérébral, dans la
gestion de l’effort.
Concernant l’effort physique, il a été démontré à de nombreuses reprises qu’il
constitue une variable de décision aversive, puisque les sujets évitent dans la mesure du
possible les options impliquant un effort physique, surtout si celui-ci est intense (Cousins and
Salamone, 1994; Hull, 1943; Salamone et al., 2007; Stevens et al., 2005; Walton et al., 2006).
Il s’agit de la loi du moindre effort formulée par Hull en 1944. La perception de l’effort par les
sujets ne repose pas sur des afférences sensitives provenant des muscles mis en jeu, mais
probablement d’une duplication de la commande motrice dans l’aire motrice supplémentaire
(White et al., 2013; Zénon et al., 2015). En conséquence, l’intensification de l’activité nerveuse
à l’origine du mouvement en cas d’administration de curares produit, pour une action
identique, une augmentation de la sensation d’effort (Gallagher et al., 2001). L’effort physique
est une notion intrinsèquement subjective, mais il peut être estimé en demandant aux sujets
de noter leur effort (Borg, 1982), en mesurant pour chaque niveau de difficulté le taux
d’acceptation des options (Jahn et al., 2020; Varazzani et al., 2015; Zénon et al., 2014), ou
encore la fréquence à laquelle les sujets acceptent de répéter les actions (Zénon et al., 2014).
Au-delà de ces mesures comportementales, plusieurs mesures physiologiques ont été
proposées : le diamètre pupillaire, la fréquence cardiaque et le taux sanguin de lactate
(Andrade et al., 2020; Da Silva et al., 2020; Tibana et al., 2019; Zénon et al., 2014). Les deux
Résumé en français
312
derniers sont sujets à des fluctuations lentes, et ne permettent donc pas d’évaluer l’effort
perçu par les sujets avec une résolution temporelle fine ou pour des actions brèves.
Concernant l’effort mental, il est défini comme la mobilisation d’énergie afin
d’atteindre un but (Gaillard, 2001, 1993), l’intensification subjective de l’activité mentale qui
relie les capacités d’un individu à effectuer une tâche à sa performance dans cette tâche (Kool
and Botvinick, 2018). Récemment, le concept d’effort mental a été unifié sous le terme de
contrôle cognitif, défini comme un ensemble de processus facilitant les comportements
dirigés vers un but (Cohen et al., 2001; Dixon and Christoff, 2012; Posner and Snyder, 2004;
Shenhav et al., 2017). Trois grandes catégories de tâches nécessitant d’exercer un effort
mental ont été identifiées : les tâches nécessitant de mettre à jour des informations stockées
dans la mémoire de travail (Kirchner, 1958), des tâches nécessitant d’exercer du contrôle
inhibiteur en particulier pour inhiber des réponses automatiques (Stroop, 1935), et enfin des
tâches nécessitant de changer régulièrement les règles appliquées (Monsell, 1996; Monsell et
al., 2003).
En ce qui concerne les mécanismes sous-jacents à l’allocation de contrôle cognitif, il a
été théorisé qu’on pouvait les décomposer en trois sous-processus (Shenhav et al., 2013). Le
premier est la régulation, qui à la modulation des processus bas-niveau de traitement
d’information. Le second est la spécification, qui se réfère aux décisions concernant
l’allocation de contrôle, définissant l’identité et l’intensité du signal de contrôle à
implémenter. Enfin, le suivi correspond à la collection d’informations concernant les
conséquences (récompenses, coûts, performance) de l’allocation de contrôle cognitif, afin
d’améliorer les futures spécifications.
Dans le contexte de la prise de décision, le contrôle cognitif est à la fois une variable à
prendre en compte, puisque les sujets s’efforcent d’éviter d’avoir à exercer du contrôle, et un
facteur d’amélioration de la prise de décision, puisque le contrôle cognitif permet d’optimiser
la prise de décision (Clark et al., 2004; Coricelli et al., 2005; De Martino et al., 2006; De Neys
et al., 2008; Eslinger and Damasio, 1985; Hinson et al., 2003; Manes et al., 2002; Mäntylä et
al., 2012; Missier et al., 2012, 2010). Ainsi, en l’absence de contrôle cognitif, les décisions sont
prises grâce à des stratégies plus simples, et probablement plus automatiques. Le contrôle
cognitif permet de faire appel à des stratégies décisionnelles plus complexes, nécessitant de
traiter plus d’informations par exemple (Gigerenzer, 2010; Gigerenzer and Goldstein, 1996;
Résumé en français
313
Sidarus et al., 2019; Simon, 1955; Tversky and Kahneman, 1974). Les sujets sont en
conséquence moins sensibles à leurs biais s’ils exercent du contrôle cognitif pour la prise de
décisions.
Concernant le rôle du contrôle cognitif comme variable de décision, il a été montré
que les sujets font un compromis entre le contrôle cognitif qu’ils investissent dans une tâche
et les bénéfices que cela peut leur rapporter (Kool et al., 2010; Kool and Botvinick, 2018, 2014;
Westbrook et al., 2013). Les sujets n’investissent généralement pas autant de contrôle cognitif
que ce dont ils seraient capables, et que ce qui permettrait de maximiser leur performanceun
constat résumé par la théorie de la rationalité limitée. Au cours des dernières années, de
nombreuses équipes ont formulé des modèles mathématiques formalisant les stratégies
d’allocation de contrôle cognitif (Griffiths et al., 2015; Lieder et al., 2014, 2012; Lieder and
Griffiths, 2017; Shenhav et al., 2017, 2013). Ces modèles se basent sur les principes
d’apprentissage par renforcement hiérarchique, c’est-à-dire qu’ils partent du principe que,
d’essai en essai, les sujets ajustent le contrôle qu’ils allouent à la tâche en fonction des
récompenses obtenues et des coûts du contrôle estimés et perçus. En d’autre termes, ils
identifient la quantité de contrôle cognitif qui leur permet de maximiser la valeur (bénéfices
moins coûts) de ce contrôle.
Puisque les sujets s’efforcent de minimiser le contrôle cognitif qu’ils allouent,
l’existence ressources qui seraient consommées pour exercer du contrôle est présumée. La
nature de ces ressources demeure hypothétique. Une première possibilité serait qu’il s’agisse
de substrats métaboliques, glucose ou glycogène (Baumeister et al., 1998; Baumeister and
Heatherton, 1996; Christie and Schrater, 2015; Muraven et al., 1998). Cette hypothèse repose
tout d’abord sur le fait que l’hypoglycémie est délétère pour l’effort mental, et la
consommation de glucose peut améliorer les performances, mais aussi sur certaines études
qui ont rapporté que les sujets voient leur glycémie diminuer après un effort mental (Benton
et al., 1987; Donohoe and Benton, 1999; Fairclough and Houston, 2004; Gailliot et al., 2007;
Kennedy and Scholey, 2000). Compte tenu que ce dernier résultat n’a pas été reproduit par la
suite, le rôle du glucose comme ressource limitante est maintenant discuté. Le glycogène, une
forme de stockage de glucose présente dans les astrocytes, a été proposé à sa place (Christie
and Schrater, 2015). Une seconde possibilité pour expliquer la dimension aversive du contrôle
cognitif repose sur l’existence d’une limite aux computations réalisables par le cerveau à un
instant donné (Allport, 1954; Anderson, 1983; Cohen et al., 1990; Logan, 1985). Toutefois,
Résumé en français
314
cette théorie ne permet d’expliquer que l’impossibilité à réaliser deux tâches cognitives en
parallèle (Shenhav et al., 2017). Enfin, une troisième possibilité purement motivationnelle
repose sur l’existence de coûts d’opportunité et sur la nécessité d’un équilibre entre
comportements d’exploitation et d’exploration (Chittka et al., 2003; Franks et al., 2003; Heitz
and Schall, 2012; Inzlicht et al., 2014; Inzlicht and Schmeichel, 2012; Wickelgren, 1977).
Compte tenu de l’impossibilité d’effectuer d’autres tâches en parallèle, allouer du contrôle à
une tâche en particulier implique de n’en effectuer aucune autre. L’allocation de contrôle
prolonge généralement les computations, ce qui veut dire qu’allouer plus de contrôle
augmente les coûts d’opportunité. En d’autres termes, l’évitement du contrôle cognitif
pourrait faciliter l’équilibre entre exploration et exploitation.
De même que l’effort physique, le contrôle cognitif est une notion subjective. Afin de
le mesurer, ou au moins de l’approximer, des mesures comportementales ont été proposées :
le temps de réaction et la performance. L’utilisation de la performance repose sur le principe
qu’à difficulté égale, augmenter le contrôle cognitif mobilisé pour la tâche augmentera la
performance (Norman and Bobrow, 1975). L’utilisation du temps de réaction est basée sur
l’idée simplificatrice que les ressources sont mobilisées avec un taux constant au cours du
temps, ce qui implique que plus de temps passé sur la tâche correspond à plus de ressources
mobilisées (Hunt et al., 2012; Krajbich et al., 2010; Krajbich and Rangel, 2011; Milosavljevic et
al., 2010; Morris et al., 2019; Philiastides and Ratcliff, 2013). Ces deux mesures
comportementales sont limitées car elles peuvent amener à un raisonnement circulaire
(Navon, 1984; Shenhav et al., 2017). En effet, il est difficile de faire la part des choses entre
des effets de la difficulté ou du contrôle cognitif mobilisé. Des mesures physiologiques ont
donc été proposées. Les principales sont les augmentations du rythme cardiaque (Carroll et
al., 1986a; Turner and Carroll, 1985) et de l’activité métabolique (Carroll et al., 1986a; Sims et
al., 1988; Sims and Carroll, 1990; Turner et al., 1983; Turner and Carroll, 1985), le diamètre
pupillaire (Alnæs et al., 2014; Hess and Polt, 1964; Kahneman and Beatty, 1966; Simpson and
Hale, 1969; Wahn et al., 2016; Wright et al., 2013), et les composants P3 et N2 des potentiels
cérébraux liés aux évènements mesurés par électroencéphalogramme (Clayson and Larson,
2013; Falkenstein et al., 1999; Folstein and Van Petten, 2008; Gajewski and Falkenstein, 2011;
Garavan et al., 1999; Jamadar et al., 2015; Kopp et al., 2020; Pratt et al., 2011; Qi et al., 2014;
Tillman and Wiens, 2011; Wei and Zhou, 2020).
Résumé en français
315
Il est intéressant de noter que malgré la catégorisation fréquente de l’effort en deux
sous catégories bien disjointes (physique et mental), il pourrait y avoir un certain
recouvrement entre ces deux concepts, et en particulier entre les ressources consommées. En
effet, faire un effort physique conduit à une diminution de la capacité pour les efforts mentaux
et vice versa (Boksem and Tops, 2008; Cadegiani and Kater, 2019; Martin et al., 2018;
McMorris, 2020; Meeusen et al., 2013; Schiphof-Godart et al., 2018; Smit, 1978; Van Cutsem
et al., 2017). Il est possible que cette relation entre les deux types d’effort vienne du fait
qu’exercer du contrôle cognitif est nécessaire à l’initiation et à la persévérance dans
l’exécution des efforts physiques, puisque le contrôle cognitif permet d’éviter les
comportements automatiques (Epstein, 1994; Sloman, 1996). Il est aussi possible que ces deux
formes d’effort consomment le même substrat métabolique : le glycogène (Christie and
Schrater, 2015; Matsui et al., 2017).
L’hypothèse de travail de cette thèse concerne l’implication du LC dans la gestion de
l’effort physique et mental. De fait, la deuxième partie de mon introduction porte sur les
caractéristiques de ce noyau noradrénergique du tronc cérébral, et sur les éléments qui ont
amené à émettre notre hypothèse. Le LC est un petit noyau dont tous les neurones produisent
de la noradrénaline (NA) (Dahlstroem and Fuxe, 1964; Goldman and Coleman, 1981;
Nieuwenhuys et al., 2008; Olszewski and Baxter, 1954; Sharma et al., 2010). C’est la seule
source de NA pour tout le cerveau antérieur. Il se situe dans le tronc cérébral, latéralement au
quatrième ventricule, ventralement à la partie la plus antérieure du cervelet, aux colliculi
inférieurs et au noyau du nerf trijumeau (Bouret and Richmond, 2009; Grant et al., 1988;
Kalwani et al., 2014). Les neurones du LC ont des caractéristiques électrophysiologiques qui
permettent de les reconnaître : un taux de décharge faible, des potentiels d’action larges, et
des réponses stéréotypées (activation puis longue inhibition) en cas de stimuli surprenants
(Bouret and Richmond, 2009; Bouret and Sara, 2004; Grant et al., 1988; Hervé-Minvielle and
Sara, 1995). Bien que les neurones du LC produisent tous de la NA, il ne s’agit pas d’un noyau
homogène : les neurones sont organisés en modules, contenant des cellules de morphologie
différente, qui produisent différents neurotransmetteurs en plus de la NA, expriment les
récepteurs à la NA en quantités différentes, reçoivent des afférences de zones cérébrales
différentes et projettent des axones vers des aires différentes (Aston-Jones et al., 1991;
Chamba et al., 1991; Léna et al., 1999; Luppi et al., 1995; Schwarz et al., 2015; Simpson et al.,
Résumé en français
316
1999; Sutin and Jacobowitz, 1991; Takeuchi et al., 2016; Xu et al., 1998; Young and Kuhar,
1980). Ajoutant à cette hétérogénéité anatomo-fonctionnelle, les effets de la NA sont
variables en fonction de l’aire cérébrale cible, car les différents récepteurs ne sont pas répartis
de manière homogène dans le cerveau, et ont des affinités différentes (Arnsten, 2000).
Compte tenu de sa localisation, l’activité du LC est difficilement mesurable, c’est
pourquoi de nombreuses équipes ont cherché à définir des mesures indirectes. Le diamètre
pupillaire (Jepma et al., 2011; Joshi et al., 2016; Murphy et al., 2014; Phillips et al., 2000;
Varazzani et al., 2015) et le composant P3 des potentiels cérébraux liés aux évènements
mesurés par électroencéphalogramme (de Rover et al., 2015; Nieuwenhuis, 2011;
Nieuwenhuis et al., 2011, 2005; Pineda et al., 1989; Swick et al., 1994; Vazey et al., 2018) sont
corrélés au taux de décharge des neurones du LC, bien qu’ils ne fournissent qu’une faible
résolution temporelle et qu’ils soient assez peu spécifiques.
Le LC et la NA sont impliqués dans plusieurs processus mobilisant des ressources. En
effet, le LC a un rôle prépondérant dans la balance entre veille et sommeil, car il promeut l’état
de veille et participe aux transitions entre sommeil et veille (Aston-Jones, 2005; Carter et al.,
2010; Foote et al., 1983; Jouvet, 1969; Saper et al., 2005; Sara, 2009; revue par Berridge and
Waterhouse, 2003). D’autre part, le LC participe aux réactions liées au stress, via des
connections à l’amygdale (Bingham et al., 2011; Bouret et al., 2003; Buffalari and Grace, 2007;
Cassens et al., 1981; Curtis et al., 2012; Koob, 1999; Sved et al., 2002; Valentino et al., 1991;
Van Bockstaele et al., 2001). Enfin, il est impliqué dans la régulation de l’attention (Bari et al.,
2020; Clark et al., 1987; Coull et al., 2001; Reynaud et al., 2019; Witte and Marrocco, 1997).
Au-delà de ces effets, le LC participe à la promotion de processus coûteux dirigés vers
un but. Il a été mentionné plus haut que la dilatation pupillaire corrèle avec le taux de
décharge du LC (Joshi et al., 2016; Murphy et al., 2014; Varazzani et al., 2015). Le diamètre
pupillaire corrèle avec quant à lui avec l’effort physique ou mental engagés dans une tâche
(Alnæs et al., 2014; Hess and Polt, 1964; Kahneman and Beatty, 1966; Simpson and Hale, 1969;
Wahn et al., 2016; Wright et al., 2013; Zénon et al., 2014). De la même manière, le composant
P3 des potentiels cérébraux liés aux évènements mesurés par électroencéphalogramme
corrèle à la fois avec l’activité du LC (de Rover et al., 2015; Nieuwenhuis, 2011; Nieuwenhuis
et al., 2011, 2005; Pineda et al., 1989; Swick et al., 1994; Vazey et al., 2018) et avec l’effort
mental (Clayson and Larson, 2013; Falkenstein et al., 1999; Folstein and Van Petten, 2008;
Gajewski and Falkenstein, 2011; Garavan et al., 1999; Jamadar et al., 2015; Kopp et al., 2020;
Résumé en français
317
Pratt et al., 2011; Qi et al., 2014; Tillman and Wiens, 2011; Wei and Zhou, 2020). Ces éléments
suggèrent que le LC pourrait participer à la gestion de l’effort. De plus, les neurones du LC
encodent la force exercée autour d’actions requérant de la force (Varazzani et al., 2015).
Par le biais de modifications pharmacologiques, chémogénétiques et optogénétiques
de l’activité du LC, il a été démontré que son activité et les taux cérébraux de NA sont
causalement impliqués dans l’exécution de tâches requérant du contrôle cognitif ou un effort
physique (Arnsten, 2006; Bari et al., 2020; Chamberlain et al., 2007, 2006a, 2006b; Coradazzi
et al., 2016; Faraone et al., 2005; Fernando et al., 2012; Gamo et al., 2010; Gilzenrat et al.,
2010; Jepma and Nieuwenhuis, 2011; Lapiz et al., 2007; Lapiz and Morilak, 2006; Li et al., 1999;
Li and Mei, 1994; Mao et al., 1999; McGaughy et al., 2008; Navarra et al., 2008; Overtoom et
al., 2003; Robinson, 2012; Robinson et al., 2008; Tait et al., 2007; Tzavara et al., 2006; Wang
et al., 2007). En particulier, nous avons démontré un rôle causal de la NA dans l’allocation
d’effort pour l’exécution d’actions nécessitant de produire de la force (Borderies et al., 2020;
Jahn et al., 2018) : suite à l’administration de clonidine, qui diminue l’activité des neurones du
LC, dans une tâche de compromis effort physique/récompense, les singes exerçaient moins
de force.
En ce qui concerne la prise de décisions, il a été mentionné plus haut que le contrôle
cognitif permet de l’optimiser. La pupille (dont le diamètre corrèle avec l’activité du LC) se
dilate au cours des décisions (Beatty, 1982; de Gee et al., 2014; Gilzenrat et al., 2010; Lempert
et al., 2015; Nassar et al., 2012). Le processus décisionnel peut se diviser en deux temps : la
sélection de l’action et le déclenchement de l’action choisie. Sélectionner une action nécessite
de traiter les informations sur les options disponibles. Les neurones du LC répondent à
l’apparition de stimuli informant les sujets sur les coûts et bénéfices des options proposées
(Bouret and Richmond, 2015; Jahn et al., 2020; Varazzani et al., 2015), d’autant plus qu’ils
fournissent des informations nouvelles (Jahn et al., 2020), ce qui serait compatible avec un
rôle de facilitation du traitement des informations. De plus, la NA a un rôle causal dans la
saillance motivationnelle, c’est-à-dire dans la régulation de l’intensité des comportements
facilitant l’accomplissement d’un objectif (Ventura et al., 2008, 2007). Par ailleurs, diminuer
l’activité du LC cause une augmentation de la sensibilité des choix à l’effort physique et une
diminution de la fréquence de sélection des options les plus difficiles (Borderies et al., 2020).
Compte tenu du biais qui consiste à éviter les efforts, engager du contrôle cognitif est
nécessaire pour choisir les options associées à un gros effort (Epstein, 1994; Sloman, 1996),
Résumé en français
318
ce qui suggère que la diminution de la fréquence des choix impliquant une action difficile
pourrait correspondre à une diminution de la capacité à engager du contrôle cognitif. En
d’autres termes, cela indiquerait que la noradrénaline pourrait être causalement impliquée
dans la mobilisation de contrôle cognitif pour la prise de décisions. L’idée du lien entre activité
du LC et contrôle cognitif est renforcée par les études qui ont montré le lien négatif entre
diamètre pupillaire et force des biais décisionnels (de Gee et al., 2020, 2017, 2014).
Pour ce qui est du déclenchement des actions choisies, les neurones du LC s’activent
généralement avant le déclenchement d’actions coûteuses et encodent la difficulté de l’action
déclenchée (Bouret et al., 2012; Bouret and Richmond, 2015, 2009; Bouret and Sara, 2004;
Clayton et al., 2004; Varazzani et al., 2015). Par ailleurs, inhiber l’activité du LC a tendance à
faire augmenter les temps de réaction (Jahn et al., 2018), qui peuvent représenter la difficulté
subjective à déclencher les actions.
Pour finir, l’hypothèse de l’implication du LC dans le suivi des conséquences des actions
et en particulier des erreurs a été émise, sur la base d’études de pupillométrie établissant un
lien entre l’amplitude des réponses pupillaires suite à une erreur et l’intensité du
ralentissement du temps de réaction suite à cette même erreur, c’est-à-dire l’intensité de
l’ajustement comportemental suite à une allocation d’effort insuffisante (Murphy et al.,
2016). De plus, des études augmentant avec peu de spécificité l’activité du LC (stimulation du
nerf vague) ont montré son lien de causalité avec le ralentissement post-erreur (Sellaro et al.,
2015).
Plusieurs mécanismes d’action de la NA, et donc de l’activité du LC pourraient
participer à son rôle dans la gestion de l’effort. Tout d’abord, la NA régule l’utilisation du
glycogène, qui a été mentionné plus haut comme potentielle ressource limitante pour
l’allocation de contrôle cognitif et d’effort physique. Dans les aires cérébrales cibles du LC, la
NA libérée facilite la glycogénolyse pour former du lactate, un substrat métabolique pour les
neurones (Magistretti, 1988; Quach et al., 1978; Sorg and Magistretti, 1991). A plus long
terme, la NA favorise la resynthèse de glycogène, en augmentant la production de protéines
critiques pour ce processus (Allaman et al., 2000; Sorg and Magistretti, 1992). Ces effets
métaboliques ont des conséquences positives sur les processus cognitifs (Gibbs, 2015; Gibbs
et al., 2010).
D’autre part, la NA facilite le traitement des informations sensorielles, puisque les
modifications électriques ou pharmacologiques de l’activité du LC augmentent le rapport
Résumé en français
319
signal/bruit, la précision temporelle des réponses, et le filtrage des stimuli dans les aires
cérébrales associées au traitement de l’information sensorielle (Ciombor et al., 1999; Devilbiss
et al., 2006; Foote et al., 1975; Halliday et al., 1994; Heath et al., 2006; Hurley et al., 2004;
Jiang et al., 1996; Kossl and Vater, 1989; Segal and Bloom, 1976; Turetsky and Fein, 2002;
Waterhouse et al., 2000, 1998). Cet effet a un intérêt particulier dans le contexte de l’effort,
car il a été mentionné plus tôt que le contrôle cognitif modifie l’identité des informations
traitées et l’intensité de leur traitement pour faciliter l’accomplissements des objectifs. Ces
effets facilitateurs du codage neuronal sont aussi présents dans des aires associées au
contrôle moteur (cervelet, Freedman et al., 1977; Hoffer et al., 1973, aire motrice primaire,
Plewnia et al., 2004) et au contrôle cognitif (cortex préfrontal, Li et al., 1999; Wang et al.,
2007).
Deux grandes théories qui donnent un contexte plus global à ces effets ont été
développées. Elles reposent sur la relation bien connue entre taux de décharge du LC et
performance cognitive, en forme de U inversé (Yerkes and Dodson, 1908). La première théorie
est celle de la régulation des états mentaux et des processus associés à ces états (Berridge and
Waterhouse, 2003). Elle formule l’idée que l’activité du LC pourrait mettre le cerveau dans un
certain état (par exemple veille ou sommeil) et qu’au sein de cet état elle pourrait moduler les
processus mentaux en cours. La théorie du gain adaptatif complémente cette première idée
(Aston-Jones and Cohen, 2005). Elle repose sur l’existence (maintenant discutée) de deux
modes d’activité du LC : tonique, avec un faible rapport signal/bruit, une activité de base
élevée et des réponses aux évènements modérées, et phasique, avec un fort rapport signal
bruit, une activité de base faible et des réponses aux évènements fortes. Le mode tonique
serait lié à la distractibilité, au désengagement des tâches et à l’exploration, tandis que le
mode phasique serait lié à l’engagement dans les tâches et à l’exploitation. L’ensemble de ces
effets serait lié à des modulations de gain, c’est-à-dire de rapport signal/bruit dans les cibles
du LC. Ces modulations de gain pourraient être responsables de variations de l’efficacité des
processus nécessaires à l’allocation d’effort.
De plus, les effets d’ajustement du gain dans les cibles du LC peuvent être prolongés
dans le temps par des effets de la NA sur la plasticité neuronale, principalement été démontrés
dans des aires sensitives et dans l’hippocampe (Bing et al., 1992, 1991; Gubits et al., 1989;
Hagena et al., 2016; Harley and Sara, 1992; Heginbotham and Dunwiddie, 1991; Hopkins and
Johnston, 1984; Lemon et al., 2009; Maity et al., 2020; Neuman and Harley, 1983; Salgado et
Résumé en français
320
al., 2012; Segal and Bloom, 1974; Stanton and Sarvey, 1985; Stone et al., 1997, 1991; Straube
et al., 2003; Walling and Harley, 2004). Notons que les effets de modulation du gain neuronal
et de facilitation de la plasticité pourraient être liés à une synergie entre glutamate et
noradrénaline (Mather et al., 2016).
Enfin, la noradrénaline modifie la connectivité à large échelle dans le cerveau (Borchert
et al., 2016; Brink et al., 2016; Coull et al., 1999; Guedj et al., 2017b; Zerbi et al., 2019), et ces
effets pourraient faciliter les transitions entre différents processus au cours d’une tâche ou
entre différentes tâches (Bouret and Sara, 2005; Corbetta et al., 2008; Sara and Bouret, 2012).
Concrètement, l’activité du LC augmente la connectivité entre les zones cérébrales impliquées
dans la tache en cours et diminue la connectivité dans les réseaux non pertinents. Cette idée
est cohérente avec les effets de facilitation de l’adaptation aux changements de règles par la
NA, et les modifications de connectivité cérébrale pourraient être une conséquence à plus
large échelle de modulations du gain neuronal, potentiellement via des modulations du
rythme Gamma (Guedj et al., 2017a).
Pour résumer, la NA facilite le traitement d’information dans les cibles du LC et ajuste
la connectivité cérébrale aux demandes de la tâche en cours. Le LC est causalement impliqué
dans des processus non dirigés vers un but qui augmentent la consommation de ressources :
éveil, stress et attention. En ce qui concerne les comportements dirigés vers un but, et en
particulier l’effort physique, les diminutions pharmacologiques de l’activité du LC ont
démontré un lien causal entre NA et difficulté subjective (coût de l’effort), une variable
computationnelle théorique qui affecte les choix et l’exécution des actions de manière
cohérente. Les neurones du LC s’activent et la pupille se dilate autour du déclenchement et
de l’exécution d’actions coûteuses. Dans le contexte de tâches de force, l’activité des
neurones et le diamètre pupillaire au cours de l’action sont corrélés positivement avec la force
requise et exécutée, bien que la dynamique exacte de ces modulations et les contributions
relatives de la difficulté objective et subjective n’aient pas été examinées en détail. En cas
d’allocation de contrôle cognitif, des mesures indirectes de l’activité du LC (P3 et diamètre
pupillaire) sont positivement corrélés à l’allocation d’effort, tandis que les modifications
pharmacologiques ou optogénétiques de l’activité du LC influencent la performance dans les
tâches impliquant du contrôle cognitif. Les neurones du LC répondent à la présentation
d’informations pertinente pour la tâche, et le diamètre pupillaire augmente au cours des
Résumé en français
321
décisions, prédisant négativement la force des biais décisionnels et donc le contrôle cognitif
alloué aux décisions. Enfin, les neurones du LC répondent aux erreurs, au moins les erreurs
d’engagement, et pourraient donc signaler les cas d’allocation insuffisante d’effort et faciliter
les ajustements comportementaux requis. En cohérence avec cette idée, les augmentations
non spécifiques de l’activité du LC ont des effets bénéfiques sur les modifications du
comportement en réponse aux erreurs.
Dans l’ensemble, ces résultats suggèrent un rôle critique de la NA et du LC dans la
gestion de l’effort physique et du contrôle cognitif. Précisément, nous avons premièrement
émis l’hypothèse que les neurones du LC modulent leur activité en fonction de l’effort alloué
au moment de l’exécution des actions requérant un effort, et signalent les conséquences de
cet effort, comme les erreurs. D’autre part, sur la base des résultats de modifications longues
de l’activité du LC, nous avons émis l’hypothèse que le LC aurait un rôle central dans la
sensibilité aux coûts de l’effort et la gestion de son allocation, puisqu’il semble faciliter les
processus nécessitant de l’effort et suivre ses coûts et conséquences, potentiellement dans le
but d’ajuster les futures décisions de faire des efforts.
Afin de confirmer cette hypothèse, plusieurs points avaient besoin d’être explorés. Il
était nécessaire d’obtenir des preuves d’un codage de l’effort au moment de l’exécution de
processus coûteux au-delà des preuves circonstancielles déjà disponibles, de généraliser ce
résultat au-delà de situations nécessitant un effort physique, et d’examiner les dynamiques
de ce codage. La première étude de ma thèse s’est focalisée sur ces aspects. Dans une seconde
étude, nous nous sommes intéressés aux conséquences de l’activation du LC sur le
comportement. Spécifiquement, nous avons examiné l’activité des neurones et son lien avec
la sensibilité à l’effort et la régulation de son allocation.
Les deux études de cette thèse ont consisté à analyser l’activité de neurones uniques
du LC (247 au total) dans des tâches de compromis effort-récompense chez des macaques
rhésus. Dans la première étude, l’activité de neurones du LC a été comparée dans trois tâches :
une de compromis Force-Récompense, une de compromis Délai-Récompense, et une tâche
de détection de signal. Dans la tâche de compromis Force-Récompense, les singes devaient
appuyer sur une poignée à un certain niveau de force pour obtenir une certaine récompense.
Le niveau de force et la récompense étaient indiqués par un indice visuel en début d’essai.
Dans la tâche de compromis Délai-Récompense, les singes devaient effectuer une action
Résumé en français
322
simple (relâcher une poignée) pour obtenir une récompense, qui serait distribuée après un
certain délai. De la même manière, délai et récompense étaient indiqués par un indice visuel
en début d’essai. Enfin, dans la tâche de détection de signal, le singe devait appuyer sur une
poignée lorsqu’un point à l’écran changeait de couleur. De fait, les deux tâches de compromis
Récompense-Délai ou Force manipulaient le contrôle cognitif nécessaire pour déclencher
l’action, puisque le comportement automatique aurait été de ne pas la réaliser. La tâche de
compromis Force-Récompense manipulait en plus l’effort physique. Enfin, les tâches de
compromis Force-Récompense et de détection de signal manipulaient les contraintes sensori-
motrices de la réponse. Afin de rechercher un codage de l’effort ou au moins de la difficulté
subjective du déclenchement et de l’exécution de l’action par les neurones du LC et de préciser
leurs dynamiques, nous nous sommes focalisés sur le moment du déclenchement et de
l’exécution des actions. Nous avons recherché des modulations de l’activité des neurones par
le temps de réaction (qui permet d’approximer la difficulté subjective et potentiellement
l’effort mental effectué pour déclencher l’action), et par la force exercée dans les deux tâches
impliquant d’appuyer sur la poignée. Nous avons découpé l’activité des neurones entre une
fenêtre de temps précédant l’action et une autre pendant l’action (pour les deux tâches
impliquant de la force, et donc une action maintenue dans le temps).
Nous avons constaté que l’activité des neurones du LC était modulée juste avant le
déclenchement des actions par le temps de réaction, uniquement lorsque le temps de réaction
reflétait des différences de difficulté à déclencher les actions (tâches de compromis Délai ou
Force-Récompense) mais pas lorsqu’il reflétait des différences sensori-motrices (tâche de
détection de signal). D’autre part, pendant l’exécution d’actions longues (force), nous avons
constaté que les neurones codaient la quantité de force appliquée sur la poignée, uniquement
lorsque les variations de force exercée étaient associées à des variations de difficulté
subjective (uniquement dans la tâche de compromis Force-Récompense). Nous en avons donc
conclu que les neurones du LC codaient la difficulté subjective des processus en cours, et
potentiellement l’effort mobilisé. La dynamique de ces modulations n’était toutefois pas
compatible avec des effets immédiats sur la mobilisation d’effort, puisque les neurones du LC
ont une vitesse de conduction lente (Aston-Jones et al., 1985). Par contre, elles pourraient
faciliter l’allocation d’effort pour les processus qui font suite à une action difficile, et servir de
signal du coût pour les structures ajustant l’allocation d’effort par la suite.
Résumé en français
323
Notons par ailleurs que l’analyse de l’activité des neurones dans la tâche de compromis
Délai-Récompense a mis en évidence plusieurs éléments. Au moment de l’apparition de
l’indice visuel, les neurones codaient la récompense promise et le délai associé, et il y avait un
codage marginalement significatif de la probabilité d’engagement dans la condition proposée,
c’est-à-dire potentiellement de la valeur subjective de l’essai. De plus, le codage de la
récompense et du délai étaient corrélés à travers les neurones, ce qui signifie que les neurones
avaient tendance à coder la valeur subjective de la récompense. D’autre part, nous avons
montré une hétérogénéité dans l’excitabilité des neurones, car ils différaient en terme de taux
de décharge de base, et ces différences étaient associées à des différences de force de
réponse aux évènements de la tâche et de modulation par les paramètres des essais (Délai et
récompense) et par les paramètres comportementaux (taux d’engagement, temps de
réaction). Ce résultat renforce l’idée de l’hétérogénéité du LC sur le plan électrophysiologique
et fonctionnel.
Dans une seconde étude, nous nous sommes focalisés sur les effets à moyen terme de
l’activation du LC. En effet, nous avons constaté dans la première étude que les neurones
codaient la difficulté subjective des processus en cours mais avec une dynamique
incompatible avec des effets directs sur ces processus. Pour cela, nous avons utilisé une tâche
de choix (2 alternatives, compromis Force-Récompense) et avons étudié l’activité de 54
neurones uniques du LC. Le singe devait d’abord appuyer sur une poignée placée devant lui
pour que deux offres, combinaisons de niveau de force à effectuer et de récompense promise,
lui soient présentées. Il devait choisir entre ces deux offres en appuyant sur une des deux
poignées placées à sa droite (option de droite) et à sa gauche (option de gauche). Nous avons
identifié plusieurs effets de l’activation du LC. Tout d’abord, pour chaque processus coûteux
(appuyer sur la poignée du milieu, faire le choix, déclencher l’action choisie, l’exécuter), les
neurones du LC s’activaient transitoirement, et cette activation reflétait généralement la
difficulté subjective du processus en cours. D’autre part, à l’échelle d’un essai, l’activation des
neurones conduisait par la suite à des temps de réaction réduits, des choix plus biaisés vers
les options difficiles, et des appuis plus forts, suggérant une diminution de la difficulté
subjective. A travers les essais, l’activité des neurones du LC au moment de l’appui choisi à
l’essai N avait pour conséquence un ajustement des choix basés sur l’effort à l’essai N+1, c’est-
à-dire qu’une activation plus forte pour l’appui d’un essai conduisait à des choix plus sensibles
Résumé en français
324
à l’effort à l’essai suivant. Enfin, nous avons mis en évidence que les neurones du LC
s’activaient après une erreur d’exécution (force insuffisante ou maintenue pendant une durée
trop faible). Nous avons constaté que cet effet pourrait en plus être associé à un codage de
l’erreur de prédiction de la récompense, même si nos résultats demandent à être confirmés.
Au bilan, de manière cohérente à travers trois taches comportementales de
compromis effort-récompense, nous avons montré que les neurones du LC modulent leur
activité en fonction de la difficulté subjective des processus en cours (déclenchement et
exécution des actions). De plus, la seconde étude a permis de préciser la relation de ces
activations avec la sensibilité à l’effort et la mobilisation d’effort physique et mental. Nous
avons montré qu’au sein d’un essai, les augmentations d’activité du LC précèdent des
diminutions de sensibilité à l’effort (augmentation de la fréquence de choix des options
difficiles, diminution du temps de réaction, augmentation de la force exercée). D’autre part,
nous avons montré que les neurones du LC répondent aux erreurs d’exécution, et que les
augmentations d’activité au moment de l’exécution des actions sont associées à des
modifications du compromis effort-récompense pour l’essai suivant, avec un poids plus
important pour l’effort.
Dans la dernière partie du manuscrit, plusieurs points sont discutés. Tout d’abord, la
dynamique des réponses du LC au cours des tâches cognitives étudiées dans le cadre ma thèse
ainsi que dans le cadre d’études réalisées dans le passé semblent compatibles avec le fait que
les neurones répondent à la fin des processus décisionnels, qu’il s’agisse de décisions de
s’engager dans des actions, de choix entre des options, ou de déclenchement d’actions
(décisions de s'engager : Bouret et al., 2012; Bouret and Richmond, 2009; Jahn et al., 2020;
Varazzani et al., 2015; déclenchement des actions: Bouret et al., 2012; Bouret and Richmond,
2015, 2009; Bouret and Sara, 2004; Clayton et al., 2004; Varazzani et al., 2015)..
Par ailleurs, le codage de la difficulté subjective par les neurones du LC, ainsi que le
potentiel codage de l’effort engagé, résonnent aves des études de pupillométrie qui ont mis
en évidence un lien entre effort physique et dilatation pupillaire (Zénon et al., 2014), et avec
des études d’imagerie calcique dans le locus coeruleus qui ont montré un lien entre l’activité
des neurones du LC et l’intensité du contrôle inhibiteur engagé dans une tâche de Posner (Bari
et al., 2020). Ce codage de l’effort pourrait provenir d’afférences du système nerveux
sympathique, et en particulier du noyau Paragigantocellularis (Aston-Jones et al., 1991, 1986;
Résumé en français
325
Ennis and Aston-Jones, 1988, 1986; Van Bockstaele and Aston-Jones, 1995), bien que leur
apparition semble tardive vu la vitesse de conduction rapide entre le noyau
Paragigantocellularis et le LC (Ennis and Aston-Jones, 1988). Il pourrait aussi provenir
d’afférences du cortex cingulaire antérieur, dont l’activité reflète l’effort et la difficulté dans
des tâches d’effort physique et mental (Croxson et al., 2009; Forstmann et al., 2006; Hillman
and Bilkey, 2010; Kennerley et al., 2009; Mars et al., 2005; Morris et al., 2019; Shenhav et al.,
2014; Walton et al., 2004; Yoshida and Ishii, 2006).
D’autre part, nous avons pu confirmer qu’au moment de la présentation d’indices
visuels informant sur les paramètres de l’essai à venir, les neurones du LC ne codent que les
paramètres de la récompense (taille et délai) et pas l’effort physique proposé (Bouret et al.,
2012; Bouret and Richmond, 2015; Jahn et al., 2020; Varazzani et al., 2015). Ces éléments
indiquent que le LC ne peut probablement pas coder la valeur subjective des options
impliquant de l’effort, mais qu’il code les paramètres de la récompense (taille, délai, et
probablement incertitude). De fait, l’activité liée aux indices visuels ne peut pas en elle-même
expliquer la relation causale entre activité du LC et les choix basés sur l’effort (Borderies et al.,
2020). Toutefois, dans la seconde étude, nous avons pu proposer un mécanisme par lequel
l’activité des neurones influencent ces choix, puisque nous avons montré que les
augmentations d’activité précédaient des diminutions de la difficulté subjective. Par ailleurs,
nous avons montré que l’activité des neurones du LC est compatible avec un suivi de l’effort
engagé et de ses conséquences (erreurs), et pourrait participer à l’ajustement des choix futurs
en fonction de l’effort engagé par le passé. Ce rôle pourrait faire intervenir les connections
entre le cortex cingulaire antérieur et le LC (Schwarz et al., 2015; Shenhav et al., 2013). D’autre
part, la diminution de la difficulté subjective des processus impliquant de l’effort pourrait être
liée aux effets de la NA sur le métabolisme des aires cérébrales cibles et/ou à des effets de
synergie entre le glutamate présent en grande quantité dans les aires les plus actives et la NA.
Les effets de promotion de l’effort et de la variabilité comportementale par le LC
(Borderies et al., 2020; Dubois et al., 2021; Jahn et al., 2018; Kane et al., 2017) pourraient
sembler indépendants, pourtant, si le LC intervient dans la régulation essai par essai de la
difficulté subjective et de l’effort engagé à l’essai suivant, les variations de l’activité des
neurones pourraient au moins en partie expliquer le bruit dans les choix.
Pour finir, nous avons montré dans la première étude que les neurones du LC sont
hétérogènes sur le plan du taux de décharge, et que le taux de décharge de base prédit la
326
force des réponses aux évènements et la force de la modulation par les paramètres de la
tâches et le comportement. Nous ne pouvons pas exclure que les effets multiples de l’activité
du LC, sur l’effort physique, mental et la variabilité dans les choix pourraient provenir de
l’activité de différents modules de ce noyau. Cette idée est cohérente avec les différents effets
observés en fonction du dosage des agents pharmacologiques utilisés pour inhiber l’activité
du LC (Borderies et al., 2020; Jahn et al., 2018). Il serait en conséquence intéressant pour de
futures études de disséquer les effets des différentes projections du LC.
327
Annex: Pharmacological evidence for the implication
of noradrenaline in effort, Borderies et al., 2020