Timezone: »
Recent work in reinforcement learning has emphasized the power of L1 regularization to perform feature selection and prevent overfitting. We propose formulating the L1 regularized linear fixed point problem as a linear complementarity problem (LCP). This formulation offers several advantages over the LARS-inspired formulation, LARS-TD. The LCP formulation allows the use of efficient off-the-shelf solvers, leads to a new uniqueness result, and can be initialized with starting points from similar problems (warm starts). We demonstrate that warm starts, as well as the efficiency of LCP solvers, can speed up policy iteration. Moreover, warm starts permit a form of modified policy iteration that can be used to approximate a "greedy" homotopy path, a generalization of the LARS-TD homotopy path that combines policy evaluation and optimization.
Author Information
Jeff Johns (US Government)
Christopher Painter-Wakefield (Duke University)
Ronald Parr (Duke University)
Related Events (a corresponding poster, oral, or spotlight)
-
2010 Poster: Linear Complementarity for Regularized Policy Evaluation and Improvement »
Wed. Dec 8th 08:00 -- 08:00 AM Room
More from the Same Authors
-
2023 Poster: A Path to Simpler Models Starts With Noise »
Lesia Semenova · Harry Chen · Ronald Parr · Cynthia Rudin -
2016 Poster: Linear Feature Encoding for Reinforcement Learning »
Zhao Song · Ronald Parr · Xuejun Liao · Lawrence Carin -
2016 Poster: Improving PAC Exploration Using the Median Of Means »
Jason Pazis · Ronald Parr · Jonathan How