Timezone: »

Introducing Symmetries to Black Box Meta Reinforcement Learning
Louis Kirsch · Sebastian Flennerhag · Hado van Hasselt · Abram Friesen · Junhyuk Oh · Yutian Chen
Event URL: https://openreview.net/forum?id=LXmP-7uObH9 »

Meta reinforcement learning (RL) attempts to discover new RL algorithms automatically from environment interaction. In so-called black-box approaches, the policy and the learning algorithm are jointly represented by a single neural network. These methods are very flexible, but they tend to underperform in terms of generalisation to new, unseen environments. In this paper, we explore the role of symmetries in meta-generalisation. We show that a recent successful meta RL approach that meta-learns an objective for backpropagation-based learning exhibits certain symmetries (specifically the reuse of the learning rule, and invariance to input and output permutations) that are not present in typical black-box meta RL systems. We hypothesise that these symmetries can play an important role in meta-generalisation. Building off recent work in black-box supervised meta learning, we develop a black-box meta RL system that exhibits these same symmetries. We show through careful experimentation that incorporating these symmetries can lead to algorithms with a greater ability to generalise to unseen action & observation spaces, tasks, and environments.

Author Information

Louis Kirsch (The Swiss AI Lab IDSIA)
Sebastian Flennerhag (DeepMind)

Ph.D. candidate in Deep Learning, focusing on network adaptation in transfer learning, meta learning and sequence learning.

Hado van Hasselt (DeepMind)
Abram Friesen (DeepMind)
Junhyuk Oh (DeepMind)
Yutian Chen (DeepMind)

More from the Same Authors