PT - JOURNAL ARTICLE AU - Seydell, Anna AU - McCann, Brian C. AU - Trommershäuser, Julia AU - Knill, David C. TI - Learning Stochastic Reward Distributions in a Speeded Pointing Task AID - 10.1523/JNEUROSCI.0647-08.2008 DP - 2008 Apr 23 TA - The Journal of Neuroscience PG - 4356--4367 VI - 28 IP - 17 4099 - http://www.jneurosci.org/content/28/17/4356.short 4100 - http://www.jneurosci.org/content/28/17/4356.full SO - J. Neurosci.2008 Apr 23; 28 AB - Recent studies have shown that humans effectively take into account task variance caused by intrinsic motor noise when planning fast hand movements. However, previous evidence suggests that humans have greater difficulty accounting for arbitrary forms of stochasticity in their environment, both in economic decision making and sensorimotor tasks. We hypothesized that humans can learn to optimize movement strategies when environmental randomness can be experienced and thus implicitly learned over several trials, especially if it mimics the kinds of randomness for which subjects might have generative models. We tested the hypothesis using a task in which subjects had to rapidly point at a target region partly covered by three stochastic penalty regions introduced as “defenders.” At movement completion, each defender jumped to a new position drawn randomly from fixed probability distributions. Subjects earned points when they hit the target, unblocked by a defender, and lost points otherwise. Results indicate that after ∼600 trials, subjects approached optimal behavior. We further tested whether subjects simply learned a set of stimulus-contingent motor plans or the statistics of defenders' movements by training subjects with one penalty distribution and then testing them on a new penalty distribution. Subjects immediately changed their strategy to achieve the same average reward as subjects who had trained with the second penalty distribution. These results indicate that subjects learned the parameters of the defenders' jump distributions and used this knowledge to optimally plan their hand movements under conditions involving stochastic rewards and penalties.