TY - JOUR
T1 - Causal model comparison shows that human representation learning is not Bayesian
AU - Geana, Andra
AU - Niv, Yael
N1 - Publisher Copyright:
© 2014 Cold Spring Harbor Laboratory Press.
PY - 2014
Y1 - 2014
N2 - How do we learn what features of our multidimensional environment are relevant in a given task? To study the computational process underlying this type of "representation learning," we propose a novel method of causal model comparison. Participants played a probabilistic learning task that required them to identify one relevant feature among several irrelevant ones. To compare between two models of this learning process, we ran each model alongside the participant during task performance, making predictions regarding the values underlying the participant's choices in real time. To test the validity of each model's predictions, we used the predicted values to try to perturb the participant's learning process: We crafted stimuli to either facilitate or hinder comparison between the most highly valued features. A model whose predictions coincide with the learned values in the participant's mind is expected to be effective in perturbing learning in this way, whereas a model whose predictions stray from the true learning process should not. Indeed, we show that in our task a reinforcement-learning model could help or hurt participants' learning, whereas a Bayesian ideal observer model could not. Beyond informing us about the notably suboptimal (but computationally more tractable) substrates of human representation learning, our manipulation suggests a sensitive method for model comparison, which allows us to change the course of people's learning in real time.
AB - How do we learn what features of our multidimensional environment are relevant in a given task? To study the computational process underlying this type of "representation learning," we propose a novel method of causal model comparison. Participants played a probabilistic learning task that required them to identify one relevant feature among several irrelevant ones. To compare between two models of this learning process, we ran each model alongside the participant during task performance, making predictions regarding the values underlying the participant's choices in real time. To test the validity of each model's predictions, we used the predicted values to try to perturb the participant's learning process: We crafted stimuli to either facilitate or hinder comparison between the most highly valued features. A model whose predictions coincide with the learned values in the participant's mind is expected to be effective in perturbing learning in this way, whereas a model whose predictions stray from the true learning process should not. Indeed, we show that in our task a reinforcement-learning model could help or hurt participants' learning, whereas a Bayesian ideal observer model could not. Beyond informing us about the notably suboptimal (but computationally more tractable) substrates of human representation learning, our manipulation suggests a sensitive method for model comparison, which allows us to change the course of people's learning in real time.
UR - http://www.scopus.com/inward/record.url?scp=84949978979&partnerID=8YFLogxK
UR - http://www.scopus.com/inward/citedby.url?scp=84949978979&partnerID=8YFLogxK
U2 - 10.1101/sqb.2014.79.024851
DO - 10.1101/sqb.2014.79.024851
M3 - Article
C2 - 25943767
AN - SCOPUS:84949978979
SN - 0091-7451
VL - 79
SP - 161
EP - 168
JO - Cold Spring Harbor symposia on quantitative biology
JF - Cold Spring Harbor symposia on quantitative biology
ER -