Timezone: »

Feature Construction for Inverse Reinforcement Learning
Sergey Levine · Zoran Popovic · Vladlen Koltun

Wed Dec 08 12:00 AM -- 12:00 AM (PST) @ None #None

The goal of inverse reinforcement learning is to find a reward function for a Markov decision process, given example traces from its optimal policy. Current IRL techniques generally rely on user-supplied features that form a concise basis for the reward. We present an algorithm that instead constructs reward features from a large collection of component features, by building logical conjunctions of those component features that are relevant to the example policy. Given example traces, the algorithm returns a reward function as well as the constructed features. The reward function can be used to recover a full, deterministic, stationary policy, and the features can be used to transplant the reward function into any novel environment on which the component features are well defined.

Author Information

Sergey Levine (UC Berkeley)
Zoran Popovic (University of Washington)
Vladlen Koltun (Adobe Research)

More from the Same Authors