Skip to yearly menu bar Skip to main content


Poster

Policy Continuation with Hindsight Inverse Dynamics

Hao Sun · Zhizhong Li · Xiaotong Liu · Bolei Zhou · Dahua Lin

East Exhibition Hall B + C #194

Keywords: [ Exploration; Reinforcement ] [ Reinforcement Learning and Planning -> Decision and Control; Reinforcement Learning and Planning ] [ Reinforcement Learning and Planning ]


Abstract:

Solving goal-oriented tasks is an important but challenging problem in reinforcement learning (RL). For such tasks, the rewards are often sparse, making it difficult to learn a policy effectively. To tackle this difficulty, we propose a new approach called Policy Continuation with Hindsight Inverse Dynamics (PCHID). This approach learns from Hindsight Inverse Dynamics based on Hindsight Experience Replay. Enabling the learning process in a self-imitated manner and thus can be trained with supervised learning. This work also extends it to multi-step settings with Policy Continuation. The proposed method is general, which can work in isolation or be combined with other on-policy and off-policy algorithms. On two multi-goal tasks GridWorld and FetchReach, PCHID significantly improves the sample efficiency as well as the final performance.

Live content is unavailable. Log in and register to view live content