Timezone: »
How do animals learn? This remains an elusive question in neuroscience. Whereas reinforcement learning often focuses on the design of algorithms that enable artificial agents to efficiently learn new tasks, here we develop a modeling framework to directly infer the empirical learning rules that animals use to acquire new behaviors. Our method efficiently infers the trial-to-trial changes in an animal’s policy, and decomposes those changes into a learning component and a noise component. Specifically, this allows us to: (i) compare different learning rules and objective functions that an animal may be using to update its policy; (ii) estimate distinct learning rates for different parameters of an animal’s policy; (iii) identify variations in learning across cohorts of animals; and (iv) uncover trial-to-trial changes that are not captured by normative learning rules. After validating our framework on simulated choice data, we applied our model to data from rats and mice learning perceptual decision-making tasks. We found that certain learning rules were far more capable of explaining trial-to-trial changes in an animal's policy. Whereas the average contribution of the conventional REINFORCE learning rule to the policy update for mice learning the International Brain Laboratory's task was just 30%, we found that adding baseline parameters allowed the learning rule to explain 92% of the animals' policy updates under our model. Intriguingly, the best-fitting learning rates and baseline values indicate that an animal's policy update, at each trial, does not occur in the direction that maximizes expected reward. Understanding how an animal transitions from chance-level to high-accuracy performance when learning a new task not only provides neuroscientists with insight into their animals, but also provides concrete examples of biological learning algorithms to the machine learning community.
Author Information
Zoe Ashwood (Princeton University)
Nicholas Roy (Princeton Neuroscience Institute)
Ji Hyun Bak (UC San Francisco)
Jonathan Pillow (Princeton University)
More from the Same Authors
-
2021 : Neural Latents Benchmark ‘21: Evaluating latent variable models of neural population activity »
Felix Pei · Joel Ye · David Zoltowski · Anqi Wu · Raeed Chowdhury · Hansem Sohn · Joseph O'Doherty · Krishna V Shenoy · Matthew Kaufman · Mark Churchland · Mehrdad Jazayeri · Lee Miller · Jonathan Pillow · Il Memming Park · Eva Dyer · Chethan Pandarinath -
2022 : Non-exchangeability in Infinite Switching Linear Dynamical Systems »
Victor Geadah · Jonathan Pillow -
2022 Panel: Panel 3A-3: Dynamic Inverse Reinforcement… & Single-phase deep learning… »
Zoe Ashwood · Will Greedy -
2022 Poster: Dynamic Inverse Reinforcement Learning for Characterizing Animal Behavior »
Zoe Ashwood · Aditi Jha · Jonathan Pillow -
2022 Poster: Extracting computational mechanisms from neural data using low-rank RNNs »
Adrian Valente · Jonathan Pillow · Srdjan Ostojic -
2020 Poster: High-contrast “gaudy” images improve the training of deep neural network models of visual cortex »
Benjamin Cowley · Jonathan Pillow -
2020 Poster: Identifying signal and noise structure in neural population activity with Gaussian process factor models »
Stephen Keeley · Mikio Aoi · Yiyi Yu · Spencer Smith · Jonathan Pillow -
2018 Poster: Scaling the Poisson GLM to massive neural datasets through polynomial approximations »
David Zoltowski · Jonathan Pillow -
2018 Poster: Efficient inference for time-varying behavior during learning »
Nicholas Roy · Ji Hyun Bak · Athena Akrami · Carlos Brody · Jonathan Pillow -
2018 Poster: Model-based targeted dimensionality reduction for neuronal population data »
Mikio Aoi · Jonathan Pillow -
2018 Poster: Power-law efficient neural codes provide general link between perceptual bias and discriminability »
Michael J Morais · Jonathan Pillow -
2018 Poster: Learning a latent manifold of odor representations from neural responses in piriform cortex »
Anqi Wu · Stan Pashkovski · Sandeep Datta · Jonathan Pillow -
2017 Poster: Gaussian process based nonlinear latent structure discovery in multivariate spike train data »
Anqi Wu · Nicholas Roy · Stephen Keeley · Jonathan Pillow -
2016 Poster: Bayesian latent structure discovery from multi-neuron recordings »
Scott Linderman · Ryan Adams · Jonathan Pillow -
2016 Poster: Adaptive optimal training of animal behavior »
Ji Hyun Bak · Jung Choi · Ilana Witten · Athena Akrami · Jonathan Pillow -
2016 Poster: A Bayesian method for reducing bias in neural representational similarity analysis »
Mingbo Cai · Nicolas W Schuck · Jonathan Pillow · Yael Niv -
2015 Poster: Convolutional spike-triggered covariance analysis for neural subunit models »
Anqi Wu · Il Memming Park · Jonathan Pillow