Reinforcement learning algorithms struggle with tasks that have complex hierarchical dependency structures. For this problem, humans usually represent the whole task in a structured way and solve it layer by layer. In this paper, we propose a novel approach called Past Data-Driven Adaptation in Hierarchical Reinforcement Learning (AdaHRL). AdaHRL leverages 'past samples' from a replay buffer to discover subgoals and construct a subgoal tree, effectively steering the agent's learning trajectory. Simultaneously, AdaHRL fine-tunes the data distribution of the entire replay buffer using a filter function, empowering adaptive learning within the agent. Experimental results demonstrate that our approach outperforms Unified Model-Free HRL Framework (UHRL) and Hindsight experience replay (HER) in tasks with complex hierarchical dependencies.
Yuhang SongJianyi WangThomas LukasiewiczZhenghua XuMai Xu
Ofir NachumShixiang GuHonglak LeeSergey Levine
Neale Van StolenSeung Hyun KimHuy TranGirish Chowdhary