Abstract
People's assessment of their ability varies in whether it is measured once following a task or sequentially via confidence judgments recorded throughout. Multiple models have been developed to predict one-off judgments of performance, which have often distinguished between peoples' biases about their general ability in a domain and their sensitivity to correctness. We propose a rational model of sequential self-assessment which allows us to make predictions about each individual separately-unlike in the one-off case which looks exclusively at the population level-and to identify, in addition to bias and sensitivity, the extent to which individuals' beliefs are responsive to their most recent evidence over the course of a task. We fit our model to data where participants solve algebraic equations and show that bias, sensitivity, and responsiveness vary meaningfully across participants.
Original language | English (US) |
---|---|
Pages | 397-403 |
Number of pages | 7 |
State | Published - 2020 |
Event | 42nd Annual Meeting of the Cognitive Science Society: Developing a Mind: Learning in Humans, Animals, and Machines, CogSci 2020 - Virtual, Online Duration: Jul 29 2020 → Aug 1 2020 |
Conference
Conference | 42nd Annual Meeting of the Cognitive Science Society: Developing a Mind: Learning in Humans, Animals, and Machines, CogSci 2020 |
---|---|
City | Virtual, Online |
Period | 7/29/20 → 8/1/20 |
All Science Journal Classification (ASJC) codes
- Artificial Intelligence
- Computer Science Applications
- Human-Computer Interaction
- Cognitive Neuroscience
Keywords
- Bayesian modeling
- Monte Carlo methods
- metacognition
- particle filter
- self-assessment