LEARNING SUBGOAL REPRESENTATIONS WITH SLOW DYNAMICS

  • Siyuan Li
  • , Lulu Zheng
  • , Jianhao Wang
  • , Chongjie Zhang

Research output: Contribution to conferencePaperpeer-review

29 Scopus citations

Abstract

In goal-conditioned Hierarchical Reinforcement Learning (HRL), a high-level policy periodically sets subgoals for a low-level policy, and the low-level policy is trained to reach those subgoals. A proper subgoal representation function, which abstracts a state space to a latent subgoal space, is crucial for effective goal-conditioned HRL, since different low-level behaviors are induced by reaching subgoals in the compressed representation space. Observing that the high-level agent operates at an abstract temporal scale, we propose a slowness objective to effectively learn the subgoal representation (i.e., the high-level action space). We provide a theoretical grounding for the slowness objective. That is, selecting slow features as the subgoal space can achieve efficient hierarchical exploration. As a result of better exploration ability, our approach significantly outperforms state-of-the-art HRL and exploration methods on a number of benchmark continuous-control tasks12. Thanks to the generality of the proposed subgoal representation learning method, empirical results also demonstrate that the learned representation and corresponding low-level policies can be transferred between distinct tasks.

Original languageEnglish
StatePublished - 2021
Event9th International Conference on Learning Representations, ICLR 2021 - Virtual, Online, Austria
Duration: May 3 2021May 7 2021

Conference

Conference9th International Conference on Learning Representations, ICLR 2021
Country/TerritoryAustria
CityVirtual, Online
Period05/3/2105/7/21

Fingerprint

Dive into the research topics of 'LEARNING SUBGOAL REPRESENTATIONS WITH SLOW DYNAMICS'. Together they form a unique fingerprint.

Cite this