June 2022 Bayesian inverse reinforcement learning for collective animal movement
Toryn L. J. Schafer, Christopher K. Wikle, Mevin B. Hooten
Author Affiliations +
Ann. Appl. Stat. 16(2): 999-1013 (June 2022). DOI: 10.1214/21-AOAS1529

Abstract

Agent-based methods allow for defining simple rules that generate complex group behaviors. The governing rules of such models are typically set a priori, and parameters are tuned from observed behavior trajectories. Instead of making simplifying assumptions across all anticipated scenarios, inverse reinforcement learning provides inference on the short-term (local) rules governing long-term behavior policies by using properties of a Markov decision process. We use the computationally efficient linearly-solvable Markov decision process to learn the local rules governing collective movement for a simulation of the selfpropelled-particle (SPP) model and a data application for a captive guppy population. The estimation of the behavioral decision costs is done in a Bayesian framework with basis function smoothing. We recover the true costs in the SPP simulation and find the guppies value collective movement more than targeted movement toward shelter.

Acknowledgments

This material is based upon work supported by the National Science Foundation Graduate Research Fellowship Program under Grant No. 1443129. Any opinions, findings, and conclusions or recommendations expressed in this material do not necessarily reflect the views of the National Science Foundation. CKW was supported by NSF Grant DMS-1811745; MBH was supported by NSF Grant DMS-1614392. Any use of trade, firm, or product names is for descriptive purposes only and does not imply endorsement by the U.S. Government.

Citation

Download Citation

Toryn L. J. Schafer. Christopher K. Wikle. Mevin B. Hooten. "Bayesian inverse reinforcement learning for collective animal movement." Ann. Appl. Stat. 16 (2) 999 - 1013, June 2022. https://doi.org/10.1214/21-AOAS1529

Information

Received: 1 September 2020; Revised: 1 June 2021; Published: June 2022
First available in Project Euclid: 13 June 2022

MathSciNet: MR4438821
zbMATH: 1498.62292
Digital Object Identifier: 10.1214/21-AOAS1529

Keywords: agent-based model , inverse optimal control , Markov decision process , variational approximation

Rights: Copyright © 2022 Institute of Mathematical Statistics

JOURNAL ARTICLE
15 PAGES

This article is only available to subscribers.
It is not available for individual sale.
+ SAVE TO MY LIBRARY

Vol.16 • No. 2 • June 2022
Back to Top