Computational modelling of the speed-accuracy tradeoff: No evidence for an association with depression symptomatology

Page created by Leon Larson
 
CONTINUE READING
Computational modelling of the speed–accuracy tradeoff: No
evidence for an association with depression symptomatology

                                James A. Grange1
                      1   School of Psychology, Keele University

                                       Abstract

    Background. Successful decision making often requires finding the right
    balance between the speed and accuracy of responding: Emphasising speed
    can lead to error-prone performance, yet emphasising accuracy leads to a
    slowing of performance. Such speed–accuracy tradeoffs (SATs) therefore
    require establishing appropriate response settings to optimise performance
    in response to changing environmental demands. Such strategic adaptaion
    of response settings relies on the striatal regions of human cortex, an area
    implicated in depression. The current study explored the association between
    depression symptomatology and SAT performance. Methods. Two exper-
    iments presented participants with an SAT paradigm embedded within a
    simple decision-making task, together with measures of depression symptoma-
    tology. Experiment 1 (N = 349) was correlational, whereas Experiment 2 was
    a two-phase experiment where participants (N = 501) were first pre-screened
    on depression symptomatology and extreme-low and extreme-high responders
    (total N = 91) were invited to Phase 2. Behavioural data were modelled with
    a drift diffusion model. Results. Behavioural data and associated diffusion
    modelling showed large and robust SAT effects. Emphasising accuracy led
    to an increase in boundary separation, an increase in drift rate, and an
    increase in non-decision time. However, the magnitude of the changes of
    these parameters with SAT instructions were not associated with measures
    of depression symptomatology. Conclusions. The results suggest that the
    strategic adaptation of response settings in response to environmental changes
    in speed–accuracy instructions do not appear to be associated with depression
    symptomatology.

    Keywords: Depression; Speed–accuracy tradeoff; Diffusion modelling
    Word count: 4,476
DEPRESSION & SPEED–ACCURACY TRADEOFFS                                                                   2

       Depression is associated with a wide range of cognitive deficits (Burt, Zembar, &
Niederehe, 1995; McDermott & Ebmeier, 2009; Rock, Roiser, Riedel, & Blackwell, 2014)
including higher-level executive functioning (Snyder, 2013) which is essential for goal-directed
behaviour. One crucial aspect of goal-directed behaviour is the ability to accurately make
rapid decisions. Many tests of cognition and executive functioning require rapid decisions
from participants in the form of choice response time (RT) tasks. For example, in the task
switching paradigms (a test of cognitive flexibility; Grange & Houghton, 2014) participants
are presented with simple stimuli (e.g., digits) and asked to rapidly switch between two
or more simple tasks (e.g., is the stimulus odd or even / is the stimulus lower or higher
than five?). Although the switching element is of primary interest to cognitive (Kiesel et al.,
2010; Vandierendonck, Liefooghe, & Verbruggen, 2010) and clinical (e.g., Ravizza & Salo,
2014) researchers, rapid decision-making is embedded within the primary task. If depression
negatively impacts cognitive processes underlying successful decision making, this could
manifest itself in tests of executive functioning (Grange & Rydon-Grange, 2021; Lawlor et
al., 2020).

       Rapid decision-making is an attractive area to study in depression and other clinical
disorders as rich theoretical developments have been made in the understanding of the
underlying cognitive processes (Ratcliff & McKoon, 2008). These theories have been devel-
oped into computational models allowing the researcher to quantify key latent parameters
reflecting the underlying cognitive processes (e.g., Voss, Nagler, & Lerche, 2013; Voss, Voss,
& Lerche, 2015). Of interest to the current study, researchers can explore the association
between these latent parameters and severity of clinical presentation (White, Ratcliff, Vasey,
& McKoon, 2010).

      The current study explores the association between depression symtpomatology and
one essential aspect of decision-making: The ability to trade speed for accuracy. Such speed–
accuracy tradeoffs (SATs)—the finding that emphasisng speed produces poorer accuracy and
emphasising accuracy slows responding—are ubiquitous in the laboratory and everyday life
(Heitz, 2014) and suggest participants can adopt different response strategies depending on
current environmental demands. The question addressed here is to establish the association
between depression symptomatology and the ability to shift response strategies.

      The paper is structured as follows. First, an overview of a popular computational
model of rapid decision-making (the drift diffusion model) is provided. I then discuss how
strategic adoption of different response strategies are accounted for by the model. I then
present the aims of the current study, and present two experiments probing the association
between depression symptomatology and shifts in response strategies.

   All raw data, analysis scripts, and code for the modelling can be downloaded from https://osf.io/48xkh/
   Correspondence concerning this article should be addressed to James A. Grange, School of Psychology,
Dorothy Hodgkin Building, Keele University, Keele, ST5 5BG, United Kingdom. E-mail: grange.jim@gmail.c
om
DEPRESSION & SPEED–ACCURACY TRADEOFFS                                                        3

The drift diffusion model (DDM)
      The DDM (see Figure 1) assumes that a decision in an RT task is the result of noisy
accumulation of sensory evidence—modelled as a diffusion process—towards one of two
response boundaries (one reflecting the correct response and the other reflecting an error
response). Evidence accumulation continues until one of the response boundaries is met.
RT of the model is determined by how long the diffusion process took to reach one of the
response boundaries, and accuracy of the model is determined by whether the diffusion
process met the correct response boundary.

Figure 1 . Schematic overview of evidence accumulation in the drift diffusion model. See
text for details. Figure available at https://tiny.cc/diffusionmodel under CC license
https://creativecommons.org/licenses/by/2.0/. Note: ms = milliseconds.

       The DDM has three main parameters. The drift rate represents the average rate of
evidence accumulation across trials, with higher values reflecting more rapid accumulation
leading to faster responding. The boundary separation reflects the height of the response
boundaries and determines how much evidence is required before a decision is made. Higher
boundary separation requires more evidence to be accumulated, prolonging RT relative to
lower boundary separation. However, lowering boundary separation increases the probability
of an error due to noise in the diffusion process. The non-decision parameter reflects elements
of the response unrelated to the decision-making process itself, such as perceptual encoding
of the stimulus and motoric responding.
       The DDM has been successfully applied to a range of clinical questions (see White et
al., 2010 for brief review), including depression. Lawlor et al. (2020) applied the DDM to
assess the impact of Major Depressive Disorder (MDD) on decision making in two studies. In
DEPRESSION & SPEED–ACCURACY TRADEOFFS                                                          4

both, the researchers found that MDD was associated with reduced drift rates and increased
boundary separation compared to controls. This extended previous work which found
impaired evidence accumulation in MDD using a model of executive control (Dillon et al.,
2015; but see Grange & Rydon-Grange, 2021). These results suggest that slowed responding
in MDD can be explained by poorer evidence accumulation in the decision-making process
together with a more cautious mode of responding.

Speed-accuracy tradeoffs (SATs)
       The higher boundary separation found in MDD could reflect the strategic adoption of a
more cautious mode of responding, but it could also reflect an inability of MDD participants
to engage in a faster (but more error-prone) mode of responding (for a similar argument
applied to healthy ageing, see Forstmann et al., 2011). This question cannot conclusively be
addressed without inducing SAT behaviour in participants. SATs can be induced in a variety
of ways (Heitz, 2014), including via explicit instructions to participants (e.g., in one block of
trials emphasise speed, and in another emphasise accuracy). The DDM accounts for SAT by
assuming an increase of the boundary separation occurs when accuracy is emphasised (cf.,
speed) thus demanding more evidence to be accumulated before committing to a response.
Studies have shown that emphasising speed over accuracy indeed leads to lower estimates of
the boundary separation parameter (see e.g., Forstmann et al., 2011; Ratcliff, Thapar, &
McKoon, 2007), but some studies have found that drift rates are also reduced (e.g., Rae,
Heathcote, Donkin, Averell, & Brown, 2014).

SATs and depression
       Strategic adaptation of response boundary settings could be disrupted because of the
involvement of the striatum component of the basal ganglia in human cortex in SAT tasks,
a cortical area also implicated in depression. Neuroimaging has shown that speed-emphasis
leads to activation of the striatum as well as the pre-supplementary motor area, and that
individual variation in activation of these structures is associated with individual variation
in estimates of boundary separation parameters of a version of the DDM (Forstmann et al.,
2008, 2011). Research has shown reduced activity in striatal regions in MDD (Hamilton
et al., 2018) which can impact goal-directed behaviour due to the interface between the
striatum and pre-frontal regions of cortex (Marquand, Haak, & Beckmann, 2017). Indeed,
depression has been shown to be associated with a deficit in functional connectivity between
striatal and pre-frontal regions (Furman, Hamilton, & Gotlib, 2011; Hamilton et al., 2018;
Pan et al., 2017).
       Given the link between the involvement of the striatum in SAT behaviour and
its disruption due to depression, examination of SAT performance—and formal DDM
modelling—could provide important insights into the cognitive profile of depression. Only
one study has examined the impact of depression on SAT performance together with DDM
application. Vallesi, Canalaz, Balestrieri, and Brambilla (2015) compared SAT performance
of 20 participants with a current or previous MDD diagnosis with 28 controls. Participants
performed a perceptual discrimination task and were cued on a trial-by-trial basis whether
to emphasise speed or accuracy in their response. Behavioural results showed the typical
SAT pattern of enhanced accuracy but longer RT on accuracy-emphasis trials, but it was
DEPRESSION & SPEED–ACCURACY TRADEOFFS                                                           5

not reported whether this behavioural pattern interacted with depression grouping. The
DDM results showed an overall reduction of drift rate in the depression group (see also
Lawlor et al., 2020). The modelling also showed that accuracy-emphasis increased estimates
of boundary separation, although again it was not reported whether this was moderated
by depression grouping. However, the results did show that those with MDD maintained a
lower boundary separation after speed-emphasis trials compared to when the previous trial
emphasised accuracy, suggesting a deficit in strategically updating response settings after
emphasising accuracy.

The current study
       The current study was designed to further examine the association between depression
and the ability to adapt response strategies to SAT instructions. Vallesi et al. (2015) cued
participants on a trial-by-trial basis whether to emphasise speed or accuracy; thus, their
study required frequent switching between response strategies, which could be disrupted by
the well-known deficit in task switching found in depression (Ravizza & Salo, 2014). Thus
it is not clear whether the findings of Vallesi et al. (2015) were caused by a deficit in the
ability to establish appropriate response strategies during SAT, or whether they were caused
by general deficits in cognitive flexibility.
      Two experiments were conducted in the current study utilising an SAT paradigm
where speed- and accuracy-emphasis occurred in separate blocks. DDM was applied to the
data to explore the magnitude of changes in model parameters across speed- and accuracy-
emphasis blocks and whether these changes were associated with the magnitude of depression
symptomatology. This design allowed me to disentangle the contribution of task switching
and the overall ability to set appropriate response strategies (cf., Vallesi et al., 2015) because
the relevant speed–accuracy emphasis was constant within blocks.
      Experiment 1 was correlational to assess whether changes in outcome variables across
speed–accuracy emphasis blocks was associated with the magnitude of depression symptoma-
tology. Experiment 2 was a between-groups design where a large number of participants
were first pre-screened on a measure of depression symtomatology, and individuals with
very-high and very-low scores were invited back to complete the SAT task.

                                        Experiment 1
Method
      The experiment was programmed and delivered online using Gorilla (Anwyl-Irvine,
Massonnié, Flitton, Kirkham, & Evershed, 2020); participants were recruited via Prolific
acadmic https://www.prolific.co/.
      Participants. Participants residing in the UK or USA were recruited; the age-range
of the sample was restricted to 18–35 because RTs are known to slow due to healthy ageing
(Salthouse, 2000). If a wider age range was included it would be hard to find a single RT
threshold for the behavioural task that would be suitably challenging for all ages.
      An a priori sensitivity analysis (Appendix A) suggested 300 participants would provide
over 86% power to detect the effect size of interest ((r = −0.18); Grange & Rydon-Grange,
DEPRESSION & SPEED–ACCURACY TRADEOFFS                                                           6

2021). 375 participants completed the study, but some were removed due to pre-specified
exclusion criteria. 2 were removed due to failing the attention check embedded within the
SHPS questionnaire. 17 were removed for failure to retain a session-wise accuracy rate
above 75% on the behavioural task. 6 were removed due to having more than 25% of their
responses slower than the RT threshold of 500ms. A further 1 participant was removed due
to an error with the experimental programme. Behavioural analysis was conducted on 349
participants. 5 participants were removed from the DDM analysis as the model did not fit
their data well.
        Materials.
        Quick inventory of depressive symptomatology–self-report (QIDS–SR).
The 16-item self-report version of the Quick Inventory of Depressive Symptomatology,
QIDS-SR-16 (Rush et al., 2003), was used. The QIDS has been shown to have excellent
internal consistency and validity (Rush et al., 2003). It consists of 16 items assessing the
severity of depressive symptoms experienced by participants in the preceding seven days.
Items relate to particular aspects of depression symptoms (e.g., “views of self”) and requires
participants to select one response from four options that best represents their experience
(e.g., “I am more self-blaming than usual”).
        Snaith–Hamilton pleasure scale (SHPS). The SHPS is a 14-item self-report
questionnaire measuring the participants’ ability to experience pleasure “. . . over the last few
days” (Snaith et al., 1995). Participants were required to read a statement (e.g., “I would
enjoy seeing other people’s smiling faces”) and indicate their agreement with this statement
on a four-point scale (Definitely Agree; Agree; Disagree; Strongly Disagree). The scoring
method recommended by Franken, Rassin, and Muris (2007) was used.
        Attention check. An attention check was embedded within the SHPS to identify
participants who were not paying attention to the items. It read “It is important that you
pay attention to this study. Please select Disagree”. Participants who did not select this
response were deemed to have failed the attention check and were removed before analysis.
        Behavioural Task. The behavioural task developed by Lawlor et al. (2020) was
utilised. The task presents participants with a schematic face within the centre of a square
frame in the centre of the screen. The participants’ task was to judge whether the mouth was
“long” or “short”. During the instruction phase, participants were informed that the “short”
mouth lies entirely within the boundaries of the eyes, whereas the “long” mouth extends to
the edge of the eyes. Each trial began with a black fixation cross presented centrally for
500ms. After this, a schematic face minus the mouth appeared for 500 ms, at which point
the mouth was presented for 100 ms. After this, the mouth disappeared, and the schematic
face remained on screen until a participant made a response. Participants were required to
press the “Z” key if the mouth was short, and the “M” key if the mouth was long.
         The behavioural task was presented in stages to help the participants learn the task,
and then to acclimatise to the speed—accuracy instructions. Participants were first given
instructions on how to perform the mouth-length task which asked participants to respond
“. . . as quickly and as accurately as possible”. After these instructions, participants completed
a 16-trial practice block. In this practice block, participants were given feedback on the
accuracy of their responses to help them learn the task. If the participant made a correct
response, a green tick was presented centrally for 200ms; if the participant made an error, a
red cross was presented centrally for 200ms.
DEPRESSION & SPEED–ACCURACY TRADEOFFS                                                                     7

       After this initial practice, participants were given instructions about the speed–accuracy
emphasis instructions. They were told that they would experience “accuracy-emphasis”
and “speed-emphasis” blocks, and that in accuracy-emphasis blocks, they should respond
“as accurately as possible without taking too long to respond”. They were told that in
speed-emphasis blocks they should respond “as quickly as possible without making too many
mistakes. DO NOT GUESS”. After these instructions, participants were presented with a
practice session consisting of 12 trials under accuracy-emphasis instructions followed by 12
trials under speed-emphasis instructions.
       Response feedback on these trials changed depending on which instructions they
were performing under. In accuracy-emphasis trials, if participants made an error they
saw the message “INCORRECT!” presented centrally in red font for 1500ms. On speed
trials, if participants responded slower than 500ms they saw the message “TOO SLOW!”
for 1500ms. 500ms was selected as an appropriate threshold to initiate speed feedback
based on performance in a pilot study (see Pilot Study 1 in Appendix B). No feedback was
provided for correct responses in accuracy-emphasis blocks, and no feedback was provided
for sub-threshold responses in speed-emphasis blocks. (See Pilot Study 2 in Appendix B for
evidence these instructions produce the desired behavioural effects.)
       After this second practice, participants were then presented with 8 blocks of 48 trials
with self-paced rest screens between each block. The relevant emphasis (speed vs. accuracy)
alternated every block, beginning with accuracy emphasis.
       Procedure. Upon entering the study online via Prolific, participants provided
informed consent, and then completed a short demographics questionnaire. After this, the
experimental software randomly allocated the participant to one of four presentation orders
of study elements: (1) behavioural task–QIDS–SHPS; (2) behavioural task–SHPS–QIDS; (3)
QIDS–SHPS–behavioural task; or (4) SHPS–QIDS–behavioural task. Upon completion of
all stages, participants were presented with a debrief screen providing detailed information
of the study.
       Quality checks and data exclusion. Participants who failed the attention checks
were removed from the analysis. Participants were also removed whose mean accuracy
performance across the experiment was below 75%. For response time analysis, error trials
were removed, and correct response times were trimmed to only include RTs longer than
150ms and shorter than 5,000ms. For the diffusion modelling, participants whose data were
not fit well by the model were removed.

Results

       All of the analysis utilised R (Version 4.0.2; R Core Team, 2020)1 ,
   1
     We, furthermore, used the R-packages brms (Version 2.13.0; Bürkner, 2017, 2018), dplyr (Version 1.0.6;
Wickham et al., 2021), forcats (Version 0.5.0; Wickham, 2020), ggplot2 (Version 3.3.3; Wickham, 2016), here
(Version 1.0.1; Müller, 2020), MASS (Version 7.3.51.6; Venables & Ripley, 2002), papaja (Version 0.1.0.9997;
Aust & Barth, 2020), purrr (Version 0.3.4; Henry & Wickham, 2020), Rcpp (Version 1.0.7; Eddelbuettel &
François, 2011; Eddelbuettel & Balamuta, 2018), readr (Version 1.3.1; Wickham, Hester, & Francois, 2018),
rtdists (Version 0.11.2; Singmann, Brown, Gretton, & Heathcote, 2020), stringr (Version 1.4.0; Wickham,
2019), tibble (Version 3.1.1; Müller & Wickham, 2021), tidyr (Version 1.1.3; Wickham, 2021), tidyverse
(Version 1.3.0; Wickham, Averick, et al., 2019), and tinylabels (Version 0.1.0; Barth, 2020).
DEPRESSION & SPEED–ACCURACY TRADEOFFS                                                         8

       Questionnaire scores. There was good variability of scores in the QIDS ques-
tionnaire (range = 0–24, M=9.08, SD=5.29) and the SHPS questionnaire (range = 14–48,
M=27.71, SD=6.21). In the QIDS, 77 participants had scores greater than or equal to 14,
and 154 had scores lower than 8, which were the thresholds used by Dillon et al. (2015)
for inclusion in their major depression and control groups, respectively. Bayesian linear
regression confirmed that QIDS scores were predicted by SHPS scores, b = 0.47, 95% credible
interval (CI; 0.39, 0.54).
       Behavioural data. For the response time analysis, error trials were removed. Cor-
rect RTs were trimmed to only include RTs longer than 150ms and shorter than 5,000ms. The
analysis of the behavioural data showed the typical effects for speed–accuracy tradeoff instruc-
tions: Proportion accuracy was significantly higher in accuracy-emphasis blocks (M=0.944,
SE=0.002) than in the speed-emphasis blocks (M=0.860, SE=0.005), t(348)=20.26, p
Table 1
Posterior estimates from the Bayesian linear regressions predicting various outcome measures from from QIDS and
SHPS scores. Note that for the beta estimates, 95% Credible Intervals are provided in square brackets.
                                                         QIDS                                    SHPS
 Source                 Outcome             Intercept             beta              Intercept             beta
 Behavioural   Accuracy                     0.904       0.000   [-0.001,   0.001]   0.899       0.000   [-0.001,   0.001]
               RT                           398         0.413   [-0.454,   1.265]   397         0.206   [-0.529,   0.950]
               delta-Accuracy               0.081       0.000   [-0.001,   0.002]   0.092       0.000   [-0.002,   0.001]
               delta-RT                     36          0.720   [-0.131,   1.549]   34          0.288   [-0.443,   1.016]

 DDM           Boundary Separation          0.774       0.001 [-0.003, 0.005]       0.793       0.000 [-0.004, 0.003]
               Drift Rate                   4.273       -0.016 [-0.039, 0.008]      4.542       -0.015 [-0.035, 0.005]
               Non-Decision Time            0.302       0.000 [0.000, 0.001]        0.298       0.000 [0.000, 0.001]
               delta-Boundary Separation    0.144       0.005 [-0.002, 0.012]       0.208       -0.001 [-0.006, 0.005]
               delta-Drift Rate             0.901       0.011 [-0.018, 0.039]       0.889       0.004 [-0.020, 0.028]
               delta-Non-Decision Time      0.039       0.000 [-0.001, 0.001]       0.037       0.000 [-0.001, 0.001]
Note. DDM = Drift diffusion model. RT = Response time.
                                                                                                                            DEPRESSION & SPEED–ACCURACY TRADEOFFS
                                                                                                                            9
DEPRESSION & SPEED–ACCURACY TRADEOFFS                                                     10

      Diffusion modelling. The diffusion model was fit to individual participant data
using the fast-dm-30 using Kolmogorov-Smirnov parameter optimisation. Boundary sepa-
ration, drift rate, and non-decision time were allowed to freely vary between speed–accuracy
emphasis conditions; all other parameters were either fixed to zero or were not free to vary
across conditions (see Voss et al., 2015). 5 participants who showed a significant deviation
between observed and model-predicted distributions were removed from inferential analysis.
For the remaining participants, the diffusion model provided a good fit to the observed data
(see Appendix C).

      Results showed that estimates of boundary separation were higher in accuracy-
emphasis blocks (M = 0.876, SE = 0.017) than in the speed-emphasis blocks (M=0.689,
SE=0.009), t(343)=10.54, p
DEPRESSION & SPEED–ACCURACY TRADEOFFS                                                                                                                                           11

                        Overall Boundary Separation                                 Overall Drift Rate                                              Overall Non−Decision Time
                                                                             10.0

                                                                                                                                             0.40
                  1.5
                                                                              7.5
Parameter Value

                                                           Parameter Value

                                                                                                                           Parameter Value
                                                                                                                                             0.35

                  1.0                                                         5.0                                                            0.30

                                                                                                                                             0.25
                                                                              2.5
                  0.5

                                                                                                                                             0.20
                         0    5    10      15   20    25                                0       5   10      15   20   25                            0    5    10      15   20   25
                                    QIDS                                                             QIDS                                                      QIDS

                        Delta Boundary Separation                                 Delta Drift Rate                                                  Delta Non−Decision Time

                   2                                                                                                                          0.4
                                                                              4

                   1
Delta Value

                                                           Delta Value

                                                                                                                           Delta Value
                                                                                                                                              0.2
                                                                              0
                   0

                                                                                                                                              0.0
                  −1                                                         −4

                  −2
                                                                                                                                             −0.2
                        0     5    10      15   20    25                            0       5       10     15    20   25                            0    5    10      15   20   25
                                   QIDS                                                             QIDS                                                       QIDS

Figure 2 . Results of the Bayesian linear regression for overall (i.e., averaged across accuracy-
and speed-emphasis blocks) diffusion model parameters (Upper Row) and delta parameter
values (Lower Row) predicted from QIDS scores. Points represent individual participant
data, and the overlaid lines represent draws from the posterior distribution of the statistical
model.

150ms and shorter than 5,000ms. For the diffusion modelling, participants whose data were
not fit well by the model were removed.

Results

      Questionnaire scores. The mean QIDS score for the low-depression group was
2.00 (SE = 0.19, range = 0–5) and for the high-depression group was 17.19 (SE = 0.44, range
= 10–22). The mean SHPS score for the low-depression group was 22.35 (SE = 0.78, range
= 14–34) and for the high-depression group was 33.69 (SE = 0.96, range = 23–52). Bayesian
linear regression confirmed that QIDS scores were predicted by SHPS scores in Phase 2,
b = 0.70, 95%-CI (0.56, 0.84). The product-moment correlation between questionnaire scores
in Phase 1 and Phase 2 for the QIDS was r=0.985.
      Behavioural data. The behavioural data is summarised in Figure 3. Two separate
mixed factorial analyses of variance (ANOVA) were performed on the RT and accuracy
DEPRESSION & SPEED–ACCURACY TRADEOFFS                                                                                                                                             12

data. The ANOVA summary table for all analyses are shown in Table 2. Accuracy-emphasis
blocks led to a significant increase in both accuracy and RT. There was no main effect of
depression group, and—critical to the current study’s aims—no interaction.
                                   Observed Response Time                                                                Observed Accuracy
                                                                                                                  0.96

                                                                              group                                                                                 group
                             420                                                     high                                                                                  high
                                                                                     low                                                                                   low
                                                                                                                  0.93
        Response Time (ms)

                                                                                            Proportion Accuracy
                             400
                                                                                                                  0.90

                                                                                                                  0.87
                             380

                                                                                                                  0.84
                                        accuracy_trials               speed_trials                                            accuracy_trials               speed_trials
                                                          Condition                                                                             Condition

Figure 3 . Observed mean response time (in milliseconds, ms) and proportion accuracy for
both low and high depression groups as a function of speed–accuracy emphasis condition.
Error bars denote one standard error around the mean.
       Diffusion modelling. Model fitting was conducted as in Experiment 1. 3 partici-
pants (two from the low-depression group and one from the high-depression group) were
removed from analyses as their data were not fit well by the model (see Appendix C). The
results of the diffusion model analysis are shown in Figure 4 and the ANOVA summary
table for all analyses are shown in Table 2. Accuracy-emphasis blocks led to a significant
increase in boundary separation, in drift rate, and in non-decision time. There was no main
effect of depression group for any of the DDM parameters, and—critically to the current
study’s aims—no interaction.

                                                                         General Discussion
      The current study examined the association between depression symptomatology and
the ability to adopt different response strategies in response to speed–accuracy tradeoff
instructions. Given the role of the striatum component of the basal ganglia in human cortex
in adapting response strategies (Forstmann et al., 2008, 2011), and given the reported
reduced activity of this cortical region in depression (Furman et al., 2011; Hamilton et al.,
2018; Pan et al., 2017), the SAT paradigm becomes an important paradigm with which to
probe the cognitive profile of depression. In addition, the drift diffusion model (Voss et al.,
2015) was used to quantify changes in latent cognitive parameters of decision-making, and
to assess whether the magnitude of changes in these parameters with SAT instructions were
associated with depression symptomatology.
      The results showed no evidence for an association between depression symtpomatology
DEPRESSION & SPEED–ACCURACY TRADEOFFS                                                                                                                                              13

                              Boundary Separation (a)                                        Drift Rate (v)                                           Non−Decision Time (t0)
                        1.1                                                            5.0                                                     0.32
                                                   group
                                                         high
                                                         low                                                                                   0.31
                        1.0
 Mean Parameter Value

                                                                Mean Parameter Value

                                                                                                                        Mean Parameter Value
                                                                                       4.5

                                                                                                                                               0.30

                        0.9

                                                                                       4.0                                                     0.29

                        0.8
                                                                                                                                               0.28
                                                                                       3.5

                        0.7
                                                                                                                                               0.27

                                 accuracy        speed                                           accuracy       speed                                    accuracy        speed
                                       Condition                                                       Condition                                               Condition
Figure 4 . Mean diffusion model parameter estimates for both low and high depression
groups as a function of speed–accuracy emphasis condition. Error bars denote one standard
error around the mean.

Table 2
ANOVA summary table for Experiment 2 for the behavioural data and parameters of the drift
diffusion model (DDM).
     Source                                        Outcome                                                    Effect                                  F-value        p           Effect Size
     Behavioural                            Accuracy                                             Depression Grouping (D)                              0.66          .418         .006
                                                                                                 SAT Condition (S)                                    176.00
DEPRESSION & SPEED–ACCURACY TRADEOFFS                                                       14

and SAT performance. This was true for both the behavioural data (i.e., changes in RT and
accuracy) and the DDM parameters: Typical behavioural SAT effects were found across
both experiments suggesting the experimental instructions were successful in inducing large
effects on response strategies; the DDM fitting (which produced excellent fits to the data)
showed that emphasising accuracy (cf., speed) led to an increase in the boundary separation
parameter, a reduction in drift rate, and an increase in non-decision time (cf., Rae et al.,
2014), but the magnitude of changes in these parameters was not associated with depression
symptom severity.
       These results appear to stand in contrast to previous research. For example, Vallesi
et al. (2015)—in a study comparing MDD patients with controls on an SAT paradigm—
found some evidence for deficits in the MDD group on updating the boundary separation
parameter immediately after speed-emphasis trials. This suggests that this group struggled
to adapt their response strategy to a more cautious mode of responding after emphasising
speed. However, because this study utilised trial-by-trial cuing of—and hence required
rapid switching between—different speed–accuracy strategies, there is the potential that
the well-known deficit in task switching found in depression (e.g., Ravizza & Salo, 2014;
Snyder, 2013) contributed to the deficit observed by Vallesi et al. (2015). That is, the
finding of Vallesi et al. (2015) could be due to problems with rapid shifting of response
strategies, a problem with general task switching, or both. The current study removed the
requirement for frequent switching between response strategies by separating speed- and
accuracy-emphasis instructions into separate blocks. A fruitful avenue for future research
would be to compare performance on within-block switching of response strategies (i.e., like
Vallesi et al., 2015) and between-block switching of response strategies (like the study) in
the same sample to explore the contribution of task switching.

Limitations
      The age-range of our samples was restricted to 18–35 so that a single response time
deadline could be used in speed-emphasis trials (see Appendix B) without unduly penalising
older adults (due to the well-documented general slowing found in older adults, Salthouse,
2000); it remains possible therefore that the lack of an association between SAT performance
and depression symptomatology on holds for this age range. Although the use of a single
global response deadline is found in many SAT procedures (see Rae et al., 2014), future
research could address the impact of this by utilising individually tailored response deadlines
for each participant via a tracking procedure (see Heitz, 2014).
      Another limitation is that the current study did not use a clinical sample; however,
both experiments showed good variability in depression symptom severity as measured by
the QIDS. In Experiment 1, 22.06% of the sample had a QIDS score of 14 or more, which
was the cutoff used by Dillon et al. (2015) as an inclusion criterion to their depression
condition. In Experiment 2, this concern was directly addressed by pre-screening a large
number of participants on depression severity and only inviting back the extreme-low and
extreme-high scorers on the QIDS to take part in the SAT paradigm. An advantage of
recruiting non-clinical samples is that the sample size of both of the studies reported here
were large. Indeed, Experiment 1 had over 86% power to detect even a small association
between depression and SAT performance (see Appendix A). Thus although the data provide
DEPRESSION & SPEED–ACCURACY TRADEOFFS                                                     15

compelling evidence for a lack of an association, it would be prudent to replicate this study
with a clinical sample
DEPRESSION & SPEED–ACCURACY TRADEOFFS                                                       16

                                        References
Anwyl-Irvine, A. L., Massonnié, J., Flitton, A., Kirkham, N., & Evershed, J. K. (2020).
      Gorilla in our midst: An online behavioral experiment builder. Behavior Research
      Methods, 52 (1), 388–407. https://doi.org/10.3758/s13428-019-01237-x
Aust, F., & Barth, M. (2020). papaja: Prepare reproducible APA journal articles with R
       Markdown. Retrieved from https://github.com/crsh/papaja
Barth, M. (2020). Tinylabels: Lightweight variable labels. Retrieved from https://CRAN.R-
       project.org/package=tinylabels
Burt, D. B., Zembar, M. J., & Niederehe, G. (1995). Depression and memory impairment: A
       meta-analysis of the association, its pattern, and specificity. Psychological Bulletin,
       117 (2), 285–305. https://doi.org/10.1037/0033-2909.117.2.285
Bürkner, P.-C. (2017). brms: An R package for Bayesian multilevel models using Stan.
      Journal of Statistical Software, 80 (1), 1–28. https://doi.org/10.18637/jss.v080.i01
Bürkner, P.-C. (2018). Advanced Bayesian multilevel modeling with the R package brms.
      The R Journal, 10 (1), 395–411. https://doi.org/10.32614/RJ-2018-017
Dillon, D. G., Wiecki, T., Pechtel, P., Webb, C., Goer, F., Murray, L., . . . Pizzagalli, D. A.
        (2015). A computational analysis of flanker interference in depression. Psychological
        Medicine, 45 (11), 2333–2344. https://doi.org/10.1017/S0033291715000276
Eddelbuettel, D., & Balamuta, J. J. (2018). Extending extitR with extitC++: A Brief
      Introduction to extitRcpp. The American Statistician, 72 (1), 28–36. https://doi.or
      g/10.1080/00031305.2017.1375990
Eddelbuettel, D., & François, R. (2011). Rcpp: Seamless R and C++ integration. Journal
      of Statistical Software, 40 (8), 1–18. https://doi.org/10.18637/jss.v040.i08
Forstmann, B. U., Dutilh, G., Brown, S., Neumann, J., Cramon, D. Y. von, Ridderinkhof, K.
       R., & Wagenmakers, E.-J. (2008). Striatum and pre-SMA facilitate decision-making
       under time pressure. Proceedings of the National Academy of Sciences, 105 (45),
      17538–17542. https://doi.org/10.1073/pnas.0805903105
Forstmann, B. U., Tittgemeyer, M., Wagenmakers, E.-J., Derrfuss, J., Imperati, D., &
      Brown, S. (2011). The speed-accuracy tradeoff in the elderly brain: A structural
      model-based approach. Journal of Neuroscience, 31 (47), 17242–17249. https:
      //doi.org/10.1523/JNEUROSCI.0309-11.2011
Franken, I. H. A., Rassin, E., & Muris, P. (2007). The assessment of anhedonia in clinical
       and non-clinical populations: Further validation of the SnaithHamilton Pleasure
       Scale (SHAPS). Journal of Affective Disorders, 99 (1), 83–89. https://doi.org/10.101
       6/j.jad.2006.08.020
Furman, D. J., Hamilton, J. P., & Gotlib, I. H. (2011). Frontostriatal functional connectivity
      in major depressive disorder. Biology of Mood & Anxiety Disorders, 1 (1), 11.
      https://doi.org/10.1186/2045-5380-1-11
DEPRESSION & SPEED–ACCURACY TRADEOFFS                                                        17

Grange, J. A., & Houghton, G. (Eds.). (2014). Task Switching and Cognitive Control. New
      York, NY: Oxford University Press.

Grange, J. A., & Rydon-Grange, M. (2021). Computational modelling of attentional
      selectivity in depression reveals perceptual deficits. Psychological Medicine, 1–10.
      https://doi.org/10.1017/S0033291720002652

Hamilton, J. P., Sacchet, M. D., Hjørnevik, T., Chin, F. T., Shen, B., Kämpe, R., . . . Gotlib,
       I. H. (2018). Striatal dopamine deficits predict reductions in striatal functional
       connectivity in major depression: A concurrent 11C-raclopride positron emission
       tomography and functional magnetic resonance imaging investigation. Translational
       Psychiatry, 8 (1), 1–10. https://doi.org/10.1038/s41398-018-0316-2

Heitz, R. P. (2014). The speed-accuracy tradeoff: History, physiology, methodology, and
       behavior. Frontiers in Neuroscience, 8. https://doi.org/10.3389/fnins.2014.00150

Henry, L., & Wickham, H. (2020). Purrr: Functional programming tools. Retrieved from
       https://CRAN.R-project.org/package=purrr

Kiesel, A., Steinhauser, M., Wendt, M., Falkenstein, M., Jost, K., Philipp, A. M., & Koch, I.
        (2010). Control and interference in task switchingA review. Psychological Bulletin,
        136 (5), 849–874. https://doi.org/10.1037/a0019842

Kruschke, J. K., & Liddell, T. M. (2018). The bayesian new statistics: Hypothesis testing, esti-
       mation, meta-analysis, and power analysis from a bayesian perspective. Psychonomic
      Bulletin & Review, 25 (1), 178–206. https://doi.org/10.3758/s13423-016-1221-4

Lawlor, V. M., Webb, C. A., Wiecki, T. V., Frank, M. J., Trivedi, M., Pizzagalli, D. A.,
       & Dillon, D. G. (2020). Dissecting the impact of depression on decision-making.
       Psychological Medicine, 50 (10), 1613–1622. https://doi.org/10.1017/S00332917190
       01570

Marquand, A. F., Haak, K. V., & Beckmann, C. F. (2017). Functional corticostriatal
      connection topographies predict goal-directed behaviour in humans. Nature Human
      Behaviour, 1 (8), 1–9. https://doi.org/10.1038/s41562-017-0146

McDermott, L. M., & Ebmeier, K. P. (2009). A meta-analysis of depression severity and
     cognitive function. Journal of Affective Disorders, 119 (1), 1–8. https://doi.org/10.1
     016/j.jad.2009.04.022

Müller, K. (2020). Here: A simpler way to find your files. Retrieved from https://CRAN.R-
        project.org/package=here

Müller, K., & Wickham, H. (2021). Tibble: Simple data frames. Retrieved from https:
       //CRAN.R-project.org/package=tibble

Pan, P. M., Sato, J. R., Salum, G. A., Rohde, L. A., Gadelha, A., Zugman, A., . . . Stringaris,
        A. (2017). Ventral striatum functional connectivity as a predictor of adolescent
        depressive disorder in a longitudinal community-based sample. American Journal of
        Psychiatry, 174 (11), 1112–1119. https://doi.org/10.1176/appi.ajp.2017.17040430
DEPRESSION & SPEED–ACCURACY TRADEOFFS                                                     18

Rae, B., Heathcote, A., Donkin, C., Averell, L., & Brown, S. (2014). The hare and the
       tortoise: Emphasizing speed can change the evidence used to make decisions. Journal
      of Experimental Psychology. Learning, Memory, and Cognition, 40 (5), 1226–1243.
       https://doi.org/10.1037/a0036801
Ratcliff, R., & McKoon, G. (2008). The diffusion decision model: Theory and data for
        two-choice decision tasks. Neural Computation, 20 (4), 873–922. https://doi.org/10.1
        162/neco.2008.12-06-420
Ratcliff, R., Thapar, A., & McKoon, G. (2007). Application of the diffusion model to
        two-choice tasks for adults 75–90 years old. Psychology and Aging, 22 (1), 56–66.
        https://doi.org/10.1037/0882-7974.22.1.56
Ratcliff, R., Thapar, A., & McKoon, G. (2010). Individual differences, aging, and IQ in
        two-choice tasks. Cognitive Psychology, 60 (3), 127–157. https://doi.org/10.1016/j.
        cogpsych.2009.09.001
Ravizza, S. M., & Salo, R. E. (2014). Task switching in psychiatric disorders. In J. A.
       Grange & G. Houghton (Eds.), Task Switching and Cognitive Control (pp. 300–331).
       New York, NY: Oxford University Press.
R Core Team. (2020). R: A language and environment for statistical computing. Vienna,
      Austria: R Foundation for Statistical Computing. Retrieved from https://www.R-
      project.org/
Rock, P. L., Roiser, J. P., Riedel, W. J., & Blackwell, A. D. (2014). Cognitive impairment in
       depression: A systematic review and meta-analysis. Psychological Medicine, 44 (10),
       2029–2040. https://doi.org/10.1017/S0033291713002535
Rush, A. J., Trivedi, M. H., Ibrahim, H. M., Carmody, T. J., Arnow, B., Klein, D. N., . . .
       Keller, M. B. (2003). The 16-Item quick inventory of depressive symptomatology
      (QIDS), clinician rating (QIDS-C), and self-report (QIDS-SR): A psychometric
       evaluation in patients with chronic major depression. Biological Psychiatry, 54 (5),
       573–583. https://doi.org/10.1016/S0006-3223(02)01866-8
Salthouse, T. A. (2000). Aging and measures of processing speed. Biological Psychology,
       54 (1), 35–54. https://doi.org/10.1016/S0301-0511(00)00052-1
Singmann, H., Brown, S., Gretton, M., & Heathcote, A. (2020). Rtdists: Response time
      distributions. Retrieved from https://CRAN.R-project.org/package=rtdists
Snaith, R. P., Hamilton, M., Morley, S., Humayan, A., Hargreaves, D., & Trigwell, P. (1995).
        A Scale for the Assessment of Hedonic Tone the SnaithHamilton Pleasure Scale. The
        British Journal of Psychiatry, 167 (1), 99–103. https://doi.org/10.1192/bjp.167.1.99
Snyder, H. R. (2013). Major depressive disorder is associated with broad impairments
       on neuropsychological measures of executive function: A meta-analysis and review.
       Psychological Bulletin, 139 (1), 81–132. https://doi.org/10.1037/a0028727
Vallesi, A., Canalaz, F., Balestrieri, M., & Brambilla, P. (2015). Modulating speed-accuracy
         strategies in major depression. Journal of Psychiatric Research, 60, 103–108. https:
DEPRESSION & SPEED–ACCURACY TRADEOFFS                                                    19

       //doi.org/10.1016/j.jpsychires.2014.09.017
Vandierendonck, A., Liefooghe, B., & Verbruggen, F. (2010). Task switching: Interplay of
       reconfiguration and interference control. Psychological Bulletin, 136 (4), 601–626.
       https://doi.org/10.1037/a0019791
Venables, W. N., & Ripley, B. D. (2002). Modern applied statistics with s (Fourth). New
      York: Springer. Retrieved from http://www.stats.ox.ac.uk/pub/MASS4
Voss, A., Nagler, M., & Lerche, V. (2013). Diffusion models in experimental psychology: A
       practical introduction. Experimental Psychology, 60 (6), 385–402. https://doi.org/10
       .1027/1618-3169/a000218
Voss, A., Voss, J., & Lerche, V. (2015). Assessing cognitive processes with diffusion
       model analyses: A tutorial based on fast-dm-30. Frontiers in Psychology, 6, 336.
       https://doi.org/10.3389/fpsyg.2015.00336
White, C. N., Ratcliff, R., Vasey, M. W., & McKoon, G. (2010). Using diffusion models to
       understand clinical disorders. Journal of Mathematical Psychology, 54 (1), 39–52.
       https://doi.org/10.1016/j.jmp.2010.01.004
Wickham, H. (2016). Ggplot2: Elegant graphics for data analysis. Springer-Verlag New
      York. Retrieved from https://ggplot2.tidyverse.org
Wickham, H. (2019). Stringr: Simple, consistent wrappers for common string operations.
      Retrieved from https://CRAN.R-project.org/package=stringr
Wickham, H. (2020). Forcats: Tools for working with categorical variables (factors). Re-
      trieved from https://CRAN.R-project.org/package=forcats
Wickham, H. (2021). Tidyr: Tidy messy data. Retrieved from https://CRAN.R-project.or
      g/package=tidyr
Wickham, H., Averick, M., Bryan, J., Chang, W., McGowan, L. D., François, R., . . . Yutani,
      H. (2019). Welcome to the tidyverse. Journal of Open Source Software, 4 (43), 1686.
      https://doi.org/10.21105/joss.01686
Wickham, H., François, R., Henry, L., & Müller, K. (2021). Dplyr: A grammar of data
      manipulation. Retrieved from https://CRAN.R-project.org/package=dplyr
Wickham, H., Hester, J., & Francois, R. (2018). Readr: Read rectangular text data. Retrieved
      from https://CRAN.R-project.org/package=readr
DEPRESSION & SPEED–ACCURACY TRADEOFFS                                                                                                                                                                  20

                                                                       Appendix A - Sample Size Planning

      To determine the sample size for Experiment 1, the procedure introduced by Kruschke
and Liddell (2018) was employed. The procedure has several steps (visualised schematically
in Figure 5) which I outline below.

       First, a minimum effect size of interest needs to be established. For the current
study, the research question is primarily interested in the relationship between depression
severity and various behavioural outcome variables and diffusion model parameters. I used
a correlation coefficient of r = −0.18 as the effect size of interest, as this is the correlation
found in our previous work (Grange & Rydon-Grange, 2021).

       Second, idealised data that conform to this effect size of interest are simulated. For
this, I generated data from 250 simulated participants by sampling from a multivariate
normal distribution. I sampled data from two measures (each standardized with mean
zero and standard deviation of one) assuming the population-level correlation between the
measures was r = −0.18. This simulated data is shown in the left panel of Figure 5.

                                Idealised Data                                                               Simulated Data                                    Predictor Posterior Distribution

                                                                                                         2
                                                                                                                                                          6
                            2
Dependent Variable Value

                                                                             Dependent Variable Value

                                                                                                         1

                                                                                                                                                          4
                                                                                                                                                density
                                                                                                         0
                            0

                                                                                                        −1
                                                                                                                                                          2

                                                                                                        −2
                           −2

                                                                                                        −3                                                0

                                −3   −2     −1        0       1    2     3                                        −2           0            2                 −0.4      −0.3       −0.2       −0.1   0.0
                                                 Predictor Value                                                          Predictor Value                                        Beta Value

Figure 5 . Schematic example of the steps applied to conduct sample size planning for
Bayesian linear regression. See text for details.

       Then, the statistical model of choice is applied to this simulated data set. The
analytical approach in the current study was a multilevel Bayesian regression with one
outcome variable and one predictor variable. The posterior distribution of this statistical
model provides a set of plausible values for the main statistical parameters of interest (in
this case, the main parameter is the slope of the predictor variable). This is shown by the
blue lines in the left panel of Figure 5; each line represents a regression slope drawn from
the posterior of the Bayesian model. Each regression line is a plausible estimate of the true
population-level relationship between the two variables.

       Then, new sample data is generated using these plausible parameter values. I generated
new samples of participants using 1,000 parameters from the posterior distribution of the
model fit to the idealised data. Each sample had N = 300 simulated participants, which was
the sample size I wished to assess the adequacy of in this analysis. For example, the red
line highlighted in the left panel of Figure 5 shows one draw from the posterior distribution;
this draw has a slope of -0.176. We would therefore simulate new data (again from the
DEPRESSION & SPEED–ACCURACY TRADEOFFS                                                       21

multivariate normal distribution) but use r = -0.176 as the correlation parameter. This new
simulated data is shown in the middle panel of Figure 5.
      With this new sample data in hand, a new Bayesian regression model is fit to the
simulated data. Plausible regression lines from the posterior prediction of this new model fit
are shown as blue lines in the middle panel of Figure 5.
       We can then assess the posterior distribution of the key parameter of interest; in the
current case, the key parameter is the slope of the regression (i.e., the beta value for the
predictor). The right panel of Figure 5 shows the posterior distribution of the predictor in
the regression fit. We are interested in whether our research goals have been met in this
posterior distribution. Specifically, we assess whether 95% of the posterior distribution is
different from zero. In the right panel of Figure 5, the 95% Credible Interval (CI) of the
posterior is shown as the horizontal black line. As can be seen, the entirety of the CI is
below zero; in this case, we would conclude that the true parameter value is non-zero. We
would therefore conclude that we have detected an effect.
        After this, a new sample from the posterior of the idealised data model fit is selected
(i.e., a new red line is chosen) to generate new sample data, and the posterior distribution
from the model fit to the new sample data is examined to see whether the 95% CI is non-zero,
and so on (for a total of 1,000 iterations). The power analysis then counts the proportion of
sample data where the CI was non-zero.
      If the intended sample size is sufficient, a true effect should be found in a large
proportion of simulated data sets. In our power analysis (assuming r = −0.18 and N = 300),
86.2% of simulated samples had a non-zero CI. Put another way, the intended sample size of
300 had a power of 86.2%.
DEPRESSION & SPEED–ACCURACY TRADEOFFS                                                      22

                              Appendix B - Pilot Studies
Pilot 1: Establishing Baseline Performance
       In speed–accuracy tradeoff experiments, researchers typically reinforce accuracy and
speed instructions with performance feedback on a trial’s accuracy or speed (respectively)
when a particular criterion is not met. For example, in accuracy-emphasis blocks researchers
will typically present feedback (e.g., “ERROR!”) when an incorrect response is made. In
speed-emphasis blocks, feedback is provided (e.g., “TOO SLOW!”) when the participant’s
response is slower than a pre-determined threshold.The question thus arises of what response
time threshold should be used.
       The purpose of the current pilot study was to establish baseline task performance in
terms of speed and accuracy on the task without speed- or accuracy-emphasis instructions;
with this baseline performance in hand, we can then use the pilot data to establish a sensible
response time threshold for speed-emphasis blocks in the main speed–accuracy tradeoff
Experiments.
       Method.
       Participants. 50 participants were recruited via Prolific Academic residing in the
UK or USA. I used pre-screening tools within Prolific Academic to limit our sample age
range to 18–35. I maintained a tight (and young) age-range because response times are
known to slow due to healthy ageing (Salthouse, 2000). If a wider age range was included in
the sample it would be hard to find a single response time threshold that would be suitably
challenging for older and younger participants alike. Participants could only take part via a
laptop or desktop computer.
       Stimuli. The same stimuli used in the behavioural task reported in the main paper
was used in this first pilot study. See Figure 6 for a schematic overview of a single trial in
this task.

Figure 6 . Schematic overview of a single trial in the behavioural task. Note that images are
not to scale.

      Procedure. After providing informed consent, the participants were presented with
full written instructions on how to perform the task. Participants then engaged in some
practice trials before moving on to the main experimental blocks. On each trial, a black
DEPRESSION & SPEED–ACCURACY TRADEOFFS                                                      23

fixation cross was presented in the centre of the frame for 500ms. After this, the schematic
face was presented without the mouth for 500ms, at which point the mouth was presented
for 100ms. After this time, the mouth disappeared, and the schematic face remained on
screen until the participant made a response. The stimulus was selected randomly on each
trial with the constraint that there were equal numbers of long- and short-mouth stimuli
within each block.
       The participants were required to press the “Z” key if they thought the presented
mouth was short, and “M” if they thought the presented mouth was long. Response
instructions asked participants to respond as quickly and as accurately as possible as soon as
the mouth appeared, and were asked to use the index finger of each hand for the responses.
Once a response was registered, the face was removed and the fixation cross for the next
trial appeared. During the practice block, participants received feedback on the accuracy of
their responses to help them learn the task. If the response was correct, a green tick was
presented on top of the schematic face; if the response was incorrect, a red “X” was shown.
Feedback was displayed for 200ms.
       Participants were presented with 24 practice trials. The main experimental section
consisted of 6 blocks of 48 trials. There was the opportunity for a self-paced break between
each block. Once all blocks were complete, participants were presented with a debrief screen
informing them of the purpose of the study.
       Quality checks and data exclusion. As the purpose of this pilot study was to
establish typical performance in this paradigm, I did not perform any quality checks or
exclude any data for the accuracy analysis. Trials in which an error was made were removed
from the response time analysis.
       Results. Density distributions of both the mean proportion accuracy and mean
response times are shown in Figure 7. Mean proportion accuracy was 0.935 (SE = 0.007),
and mean response time was 453ms (SE = 11ms).
       Establishing a suitable threshold. Examination of the response time density
distribution in Figure 7 suggests that the majority of response times fell below 500ms.
Indeed, the mean percentage of trials that fell below 500ms was 76.7% (SE = 2.8%, min =
1.1%, max = 99.6%), suggesting that for the majority of participants 500ms would prove a
suitable challenge for a response time threshold.

Pilot 2: SAT Manipulation Check
      In this pilot study I introduced speed–accuracy manipulations by emphasising speed
and accuracy in separate blocks; this provided an opportunity to test whether the response
time threshold established in pilot study 1 was suitable. I also fitted the diffusion model to
the data.
      Success of this pilot study would be determined by being able to successfully in-
duce speed–accuracy tradeoffs with our method (that is, by observing better accuracy in
“accuracy-emphasis” blocks, and shorter RTs in “speed-emphasis” blocks). Establishing
which diffusion model parameters are affected by speed–accuracy tradeoffs would also help
establish predictions for the main experiments.
      Method.
DEPRESSION & SPEED–ACCURACY TRADEOFFS                                                                                24

               Observed Accuracy                                                   Observed Response Times
          12

                                                                           0.006
          9

                                                                           0.004
density

                                                                 density
          6

                                                                           0.002
          3

          0                                                                0.000

                    0.80        0.85     0.90      0.95   1.00                          400       500        600    700
                           Mean Accuracy (Proportion)                                     Mean Response Time (ms)
Figure 7 . Density distributions of the mean proportion accuracy and mean response times
from Pilot Study 1. ms = milliseconds.

       Participants. 50 participants aged between 18–35 (inclusive) were recruited from
Prolific Academic residing in the UK or USA. The same exclusion criteria used in Pilot
Study 1 were used. Participants who took part in Pilot Study 1 were not eligible to take
part in this study.
       Stimuli. The same stimuli from Pilot Study 1 were used.
       Procedure. The task requirements were the same as in Pilot Study 1. However,
there were some general changes to the procedure in terms of instructions. I also added an
attention check (described below) to make sure participants were reading the instructions; I
removed participants who failed this attention check from the analysis.

        When participants first entered the study, they were given general instructions that
initially did not mention speed–accuracy instructions. Instead, they were given general
instructions to learn the response requirements, and were then presented with a practice
block to learn the task. During this initial practice block, the instructions asked participants
to respond “as quickly and as accurately as possible”. Feedback was presented if participants
made an error in the same way as in Pilot Study 1. This initial practice consisted of 16
trials.

      After this, participants were given instructions about the speed–accuracy emphasis
instructions. They were told that in the main study they would experience “accuracy-
emphasis” and “speed-emphasis” blocks. They were told that in accuracy-emphasis blocks,
they should respond “. . . as accurately as possible without taking too long to respond”. They
were told that in speed-emphasis blocks they should respond “. . . as quickly as possible
without making too many mistakes. DO NOT GUESS”. After these instructions, participants
were presented with a practice session where they were presented with 12 trials under
accuracy-emphasis instructions followed by 12 trials under speed-emphasis instructions.
You can also read