Timezone: »
Understanding decision-making is a core goal in both neuroscience and psychology, and computational models have often been helpful in the pursuit of this goal. While many models have been developed for characterizing behavior in binary decision-making and bandit tasks, comparatively little work has focused on animal decision-making in more complex tasks, such as navigation through a maze. Inverse reinforcement learning (IRL) is a promising approach for understanding such behavior, as it aims to infer the unknown reward function of an agent from its observed trajectories through state space. However, IRL has yet to be widely applied in neuroscience. One potential reason for this is that existing IRL frameworks assume that an agent's reward function is fixed over time. To address this shortcoming, we introduce dynamic inverse reinforcement learning (DIRL), a novel IRL framework that allows for time-varying intrinsic rewards. Our method parametrizes the unknown reward function as a time-varying linear combination of spatial reward maps (which we refer to as "goal maps"). We develop an efficient inference method for recovering this dynamic reward function from behavioral data. We demonstrate DIRL in simulated experiments and then apply it to a dataset of mice exploring a labyrinth. Our method returns interpretable reward functions for two separate cohorts of mice, and provides a novel characterization of exploratory behavior. We expect DIRL to have broad applicability in neuroscience, and to facilitate the design of biologically-inspired reward functions for training artificial agents.
Author Information
Zoe Ashwood (Princeton University/DeepMind)
Aditi Jha (Princeton University)
Jonathan Pillow (Princeton University)
More from the Same Authors
-
2021 : Neural Latents Benchmark ‘21: Evaluating latent variable models of neural population activity »
Felix Pei · Joel Ye · David Zoltowski · Anqi Wu · Raeed Chowdhury · Hansem Sohn · Joseph O'Doherty · Krishna V Shenoy · Matthew Kaufman · Mark Churchland · Mehrdad Jazayeri · Lee Miller · Jonathan Pillow · Il Memming Park · Eva Dyer · Chethan Pandarinath -
2022 : Non-exchangeability in Infinite Switching Linear Dynamical Systems »
Victor Geadah · Jonathan Pillow -
2022 Panel: Panel 3A-3: Dynamic Inverse Reinforcement… & Single-phase deep learning… »
Zoe Ashwood · Will Greedy -
2022 Poster: Extracting computational mechanisms from neural data using low-rank RNNs »
Adrian Valente · Jonathan Pillow · Srdjan Ostojic -
2020 Poster: High-contrast “gaudy” images improve the training of deep neural network models of visual cortex »
Benjamin Cowley · Jonathan Pillow -
2020 Poster: Identifying signal and noise structure in neural population activity with Gaussian process factor models »
Stephen Keeley · Mikio Aoi · Yiyi Yu · Spencer Smith · Jonathan Pillow -
2020 Poster: Inferring learning rules from animal decision-making »
Zoe Ashwood · Nicholas Roy · Ji Hyun Bak · Jonathan Pillow -
2018 Poster: Scaling the Poisson GLM to massive neural datasets through polynomial approximations »
David Zoltowski · Jonathan Pillow -
2018 Poster: Efficient inference for time-varying behavior during learning »
Nicholas Roy · Ji Hyun Bak · Athena Akrami · Carlos Brody · Jonathan Pillow -
2018 Poster: Model-based targeted dimensionality reduction for neuronal population data »
Mikio Aoi · Jonathan Pillow -
2018 Poster: Power-law efficient neural codes provide general link between perceptual bias and discriminability »
Michael J Morais · Jonathan Pillow -
2018 Poster: Learning a latent manifold of odor representations from neural responses in piriform cortex »
Anqi Wu · Stan Pashkovski · Sandeep Datta · Jonathan Pillow -
2017 Poster: Gaussian process based nonlinear latent structure discovery in multivariate spike train data »
Anqi Wu · Nicholas Roy · Stephen Keeley · Jonathan Pillow -
2016 Poster: Bayesian latent structure discovery from multi-neuron recordings »
Scott Linderman · Ryan Adams · Jonathan Pillow -
2016 Poster: Adaptive optimal training of animal behavior »
Ji Hyun Bak · Jung Choi · Ilana Witten · Athena Akrami · Jonathan Pillow -
2016 Poster: A Bayesian method for reducing bias in neural representational similarity analysis »
Mingbo Cai · Nicolas W Schuck · Jonathan Pillow · Yael Niv -
2015 Poster: Convolutional spike-triggered covariance analysis for neural subunit models »
Anqi Wu · Il Memming Park · Jonathan Pillow