A simple model for learning in volatile environments.

Sound principles of statistical inference dictate that uncertainty shapes learning. In this work, we revisit the question of learning in volatile environments, in which both the first and second-order statistics of observations dynamically evolve over time. We propose a new model, the volatile Kalma...

Full description

Bibliographic Details
Main Authors: Payam Piray, Nathaniel D Daw
Format: Article
Language:English
Published: Public Library of Science (PLoS) 2020-07-01
Series:PLoS Computational Biology
Online Access:https://doi.org/10.1371/journal.pcbi.1007963
id doaj-f2d6e1a3070e44059e70b232d4b3a04e
record_format Article
spelling doaj-f2d6e1a3070e44059e70b232d4b3a04e2021-04-21T15:16:14ZengPublic Library of Science (PLoS)PLoS Computational Biology1553-734X1553-73582020-07-01167e100796310.1371/journal.pcbi.1007963A simple model for learning in volatile environments.Payam PirayNathaniel D DawSound principles of statistical inference dictate that uncertainty shapes learning. In this work, we revisit the question of learning in volatile environments, in which both the first and second-order statistics of observations dynamically evolve over time. We propose a new model, the volatile Kalman filter (VKF), which is based on a tractable state-space model of uncertainty and extends the Kalman filter algorithm to volatile environments. The proposed model is algorithmically simple and encompasses the Kalman filter as a special case. Specifically, in addition to the error-correcting rule of Kalman filter for learning observations, the VKF learns volatility according to a second error-correcting rule. These dual updates echo and contextualize classical psychological models of learning, in particular hybrid accounts of Pearce-Hall and Rescorla-Wagner. At the computational level, compared with existing models, the VKF gives up some flexibility in the generative model to enable a more faithful approximation to exact inference. When fit to empirical data, the VKF is better behaved than alternatives and better captures human choice data in two independent datasets of probabilistic learning tasks. The proposed model provides a coherent account of learning in stable or volatile environments and has implications for decision neuroscience research.https://doi.org/10.1371/journal.pcbi.1007963
collection DOAJ
language English
format Article
sources DOAJ
author Payam Piray
Nathaniel D Daw
spellingShingle Payam Piray
Nathaniel D Daw
A simple model for learning in volatile environments.
PLoS Computational Biology
author_facet Payam Piray
Nathaniel D Daw
author_sort Payam Piray
title A simple model for learning in volatile environments.
title_short A simple model for learning in volatile environments.
title_full A simple model for learning in volatile environments.
title_fullStr A simple model for learning in volatile environments.
title_full_unstemmed A simple model for learning in volatile environments.
title_sort simple model for learning in volatile environments.
publisher Public Library of Science (PLoS)
series PLoS Computational Biology
issn 1553-734X
1553-7358
publishDate 2020-07-01
description Sound principles of statistical inference dictate that uncertainty shapes learning. In this work, we revisit the question of learning in volatile environments, in which both the first and second-order statistics of observations dynamically evolve over time. We propose a new model, the volatile Kalman filter (VKF), which is based on a tractable state-space model of uncertainty and extends the Kalman filter algorithm to volatile environments. The proposed model is algorithmically simple and encompasses the Kalman filter as a special case. Specifically, in addition to the error-correcting rule of Kalman filter for learning observations, the VKF learns volatility according to a second error-correcting rule. These dual updates echo and contextualize classical psychological models of learning, in particular hybrid accounts of Pearce-Hall and Rescorla-Wagner. At the computational level, compared with existing models, the VKF gives up some flexibility in the generative model to enable a more faithful approximation to exact inference. When fit to empirical data, the VKF is better behaved than alternatives and better captures human choice data in two independent datasets of probabilistic learning tasks. The proposed model provides a coherent account of learning in stable or volatile environments and has implications for decision neuroscience research.
url https://doi.org/10.1371/journal.pcbi.1007963
work_keys_str_mv AT payampiray asimplemodelforlearninginvolatileenvironments
AT nathanielddaw asimplemodelforlearninginvolatileenvironments
AT payampiray simplemodelforlearninginvolatileenvironments
AT nathanielddaw simplemodelforlearninginvolatileenvironments
_version_ 1714667587696066560