TY - JOUR
T1 - Dynamic estimation of task-relevant variance in movement under risk
AU - Landy, Michael S.
AU - Trommershäuser, Julia
AU - Daw, Nathaniel D.
PY - 2012/9/12
Y1 - 2012/9/12
N2 - Humanstake into account theirownmovement variability as well as potential consequences of different movement outcomes in planning movement trajectories. When variability increases, planned movements are altered so as to optimize expected consequences of the movement. Past research has focused on the steady-state responses to changing conditions of movement under risk. Here, we study the dynamics of such strategy adjustment in a visuomotor decision task in which subjects reach toward a display with regions that lead to rewards and penalties, under conditions of changing uncertainty. In typical reinforcement learning tasks, subjects should base subsequent strategy by computing an estimate of the mean outcome (e.g., reward) in recent trials. In contrast, in our task, strategy should be based on a dynamic estimate of recent outcome uncertainty (i.e., squared error). We find that subjects respond to increased movement uncertainty by aiming movements more conservatively with respect to penalty regions, and that the estimate of uncertainty they use is well characterized by a weighted average of recent squared errors, with higher weights given to more recent trials.
AB - Humanstake into account theirownmovement variability as well as potential consequences of different movement outcomes in planning movement trajectories. When variability increases, planned movements are altered so as to optimize expected consequences of the movement. Past research has focused on the steady-state responses to changing conditions of movement under risk. Here, we study the dynamics of such strategy adjustment in a visuomotor decision task in which subjects reach toward a display with regions that lead to rewards and penalties, under conditions of changing uncertainty. In typical reinforcement learning tasks, subjects should base subsequent strategy by computing an estimate of the mean outcome (e.g., reward) in recent trials. In contrast, in our task, strategy should be based on a dynamic estimate of recent outcome uncertainty (i.e., squared error). We find that subjects respond to increased movement uncertainty by aiming movements more conservatively with respect to penalty regions, and that the estimate of uncertainty they use is well characterized by a weighted average of recent squared errors, with higher weights given to more recent trials.
UR - http://www.scopus.com/inward/record.url?scp=84866252910&partnerID=8YFLogxK
UR - http://www.scopus.com/inward/citedby.url?scp=84866252910&partnerID=8YFLogxK
U2 - 10.1523/JNEUROSCI.6160-11.2012
DO - 10.1523/JNEUROSCI.6160-11.2012
M3 - Article
C2 - 22972994
AN - SCOPUS:84866252910
SN - 0270-6474
VL - 32
SP - 12702
EP - 12711
JO - Journal of Neuroscience
JF - Journal of Neuroscience
IS - 37
ER -