Skip to yearly menu bar Skip to main content


Poster

Regret Minimization in MDPs with Options without Prior Knowledge

Ronan Fruit · Matteo Pirotta · Alessandro Lazaric · Emma Brunskill

Pacific Ballroom #14

Keywords: [ Reinforcement Learning and Planning ] [ Reinforcement Learning ] [ Model-Based RL ] [ Hierarchical RL ] [ Exploration ]


Abstract:

The option framework integrates temporal abstraction into the reinforcement learning model through the introduction of macro-actions (i.e., options). Recent works leveraged on the mapping of Markov decision processes (MDPs) with options to semi-MDPs (SMDPs) and introduced SMDP-versions of exploration-exploitation algorithms (e.g., RMAX-SMDP and UCRL-SMDP) to analyze the impact of options on the learning performance. Nonetheless, the PAC-SMDP sample complexity of RMAX-SMDP can hardly be translated into equivalent PAC-MDP theoretical guarantees, while UCRL-SMDP requires prior knowledge of the parameters characterizing the distributions of the cumulative reward and duration of each option, which are hardly available in practice. In this paper, we remove this limitation by combining the SMDP view together with the inner Markov structure of options into a novel algorithm whose regret performance matches UCRL-SMDP's up to an additive regret term. We show scenarios where this term is negligible and the advantage of temporal abstraction is preserved. We also report preliminary empirical result supporting the theoretical findings.

Live content is unavailable. Log in and register to view live content