Computational modeling of behavioral tasks: An illustration on a classic reinforcement learning paradigm

There has been a growing interest among psychologists, psychiatrists and neuroscientists in applying computational modeling to behavioral data to understand animal and human behavior. Such approaches can be daunting for those without experience. This paper presents a step-by-step tutorial to conduct...

Full description

Bibliographic Details
Main Authors: Suthaharan, Praveen, Corlett, Philip R., Ang, Yuen-Siang
Format: Article
Language:English
Published: Université d'Ottawa 2021-06-01
Series:Tutorials in Quantitative Methods for Psychology
Subjects:
r
Online Access:https://www.tqmp.org/RegularArticles/vol17-2/p105/p105.pdf
Description
Summary:There has been a growing interest among psychologists, psychiatrists and neuroscientists in applying computational modeling to behavioral data to understand animal and human behavior. Such approaches can be daunting for those without experience. This paper presents a step-by-step tutorial to conduct parameter estimation in R via three techniques: Maximum Likelihood Estimation (MLE), Maximum A Posteriori (MAP) and Expectation-Maximization with Laplace approximation (EML). We first demonstrate how to simulate a classic reinforcement learning paradigm -- the two-armed bandit task -- for N = 100 subjects; and then explain how to develop the computational model and implement the MLE, MAP and EML methods to recover the parameters. By presenting a sufficiently detailed walkthrough on a familiar behavioral task, we hope this tutorial could benefit readers interested in applying parameter estimation methods in their own research.
ISSN:1913-4126