13 research outputs found
Expressions for Bayesian confidence of drift diffusion observers in fluctuating stimuli tasks
We introduce a new approach to modelling decision confidence, with the aim of enabling computationally cheap predictions while taking into account, and thereby exploiting, trial-by-trial variability in stochastically fluctuating stimuli. Using the framework of the drift diffusion model of decision making, along with time-dependent thresholds and the idea of a Bayesian confidence readout, we derive expressions for the probability distribution over confidence reports. In line with current models of confidence, the derivations allow for the accumulation of “pipeline” evidence that has been received but not processed by the time of response, the effect of drift rate variability, and metacognitive noise. The expressions are valid for stimuli that change over the course of a trial with normally-distributed fluctuations in the evidence they provide. A number of approximations are made to arrive at the final expressions, and we test all approximations via simulation. The derived expressions contain only a small number of standard functions, and require evaluating only once per trial, making trial-by-trial modelling of confidence data in stochastically fluctuating stimuli tasks more feasible. We conclude by using the expressions to gain insight into the confidence of optimal observers, and empirically observed patterns
Bayesian confidence in optimal decisions
The optimal way to make decisions in many circumstances is to track the difference in evidence collected in
favour of the options. The drift diffusion model (DDM) implements this approach, and provides an excellent
account of decisions and response times. However, existing DDM-based models of confidence exhibit certain
deficits, and many theories of confidence have used alternative, non-optimal models of decisions. Motivated by
the historical success of the DDM, we ask whether simple extensions to this framework might allow it to better
account for confidence. Motivated by the idea that the brain will not duplicate representations of evidence, in
all model variants decisions and confidence are based on the same evidence accumulation process. We compare
the models to benchmark results, and successfully apply 4 qualitative tests concerning the relationships between
confidence, evidence, and time, in a new preregistered study. Using computationally cheap expressions to model
confidence on a trial-by-trial basis, we find that a subset of model variants also provide a very good to excellent
account of precise quantitative effects observed in confidence data. Specifically, our results favour the hypothesis
that confidence reflects the strength of accumulated evidence penalised by the time taken to reach the decision
(Bayesian readout), with the penalty applied not perfectly calibrated to the specific task context. These results
suggest there is no need to abandon the DDM or single accumulator models to successfully account for confidence
reports
The Confidence Database
Understanding how people rate their confidence is critical for the characterization of a wide range of perceptual, memory, motor and cognitive processes. To enable the continued exploration of these processes, we created a large database of confidence studies spanning a broad set of paradigms, participant populations and fields of study. The data from each study are structured in a common, easy-to-use format that can be easily imported and analysed using multiple software packages. Each dataset is accompanied by an explanation regarding the nature of the collected data. At the time of publication, the Confidence Database (which is available at https://osf.io/s46pr/) contained 145 datasets with data from more than 8,700 participants and almost 4 million trials. The database will remain open for new submissions indefinitely and is expected to continue to grow. Here we show the usefulness of this large collection of datasets in four different analyses that provide precise estimations of several foundational confidence-related effects
Explaining the effects of distractor statistics in visual search
Data and code for "Explaining the effects of distractor statistics in visual search"
Bayesian confidence in optimal decisions
The optimal way to make decisions in many circumstances is to track the difference in evidence collected in favour of the options. The drift diffusion model (DDM) implements this approach, and provides an excellent account of decisions and response times. However, existing DDM-based models of confidence exhibit certain deficits, and many theories of confidence have used alternative, non-optimal models of decisions. Motivated by the historical success of the DDM, we ask whether simple extensions to this framework might allow it to better account for confidence. Motivated by the idea that the brain will not duplicate representations of evidence, in all model variants decisions and confidence are based on the same evidence accumulation process. We compare the models to benchmark results, and successfully apply 4 qualitative tests concerning the relationships between confidence, evidence, and time, in a new preregistered study. Using computationally cheap expressions to model confidence on a trial-by-trial basis, we find that a subset of model variants also provide a very good to excellent account of precise quantitative effects observed in confidence data. Specifically, our results favour the hypothesis that confidence reflects the strength of accumulated evidence penalised by the time taken to reach the decision (Bayesian readout), with the penalty applied not perfectly calibrated to the specific task context. These results suggest there is no need to abandon the DDM or single accumulator models to successfully account for confidence reports
Exploring rapidly changing sensory-motor mappings in behaviour and MEG
*Joint first authors
Behavioural, MEG and eye tracking data from an experiment exploring how humans achieve rapid switches between arbitrary stimulus to response mappings. The experimental procedure is described in one of the preregistrations associated with the study (https://osf.io/hnwfr)
Ma Lab Resources
This is the overview project for the resources from Weiji Ma's lab at NYU