JOURNAL ARTICLE

Data-Efficient Hierarchical Reinforcement Learning

Ofir NachumShixiang GuHonglak LeeSergey Levine

Year: 2018 Journal:   arXiv (Cornell University) Vol: 31 Pages: 3303-3313   Publisher: Cornell University

Abstract

Hierarchical reinforcement learning (HRL) is a promising approach to extend traditional reinforcement learning (RL) methods to solve more complex tasks. Yet, the majority of current HRL methods require careful task-specific design and on-policy training, making them difficult to apply in real-world scenarios. In this paper, we study how we can develop HRL algorithms that are general, in that they do not make onerous additional assumptions beyond standard RL algorithms, and efficient, in the sense that they can be used with modest numbers of interaction samples, making them suitable for real-world problems such as robotic control. For generality, we develop a scheme where lower-level controllers are supervised with goals that are learned and proposed automatically by the higher-level controllers. To address efficiency, we propose to use off-policy experience for both higher and lower-level training. This poses a considerable challenge, since changes to the lower-level behaviors change the action space for the higher-level policy, and we introduce an off-policy correction to remedy this challenge. This allows us to take advantage of recent advances in off-policy model-free RL to learn both higher- and lower-level policies using substantially fewer environment interactions than on-policy algorithms. We term the resulting HRL agent HIRO and find that it is generally applicable and highly sample-efficient. Our experiments show that HIRO can be used to learn highly complex behaviors for simulated robots, such as pushing objects and utilizing them to reach target locations, learning from only a few million samples, equivalent to a few days of real-time interaction. In comparisons with a number of prior HRL methods, we find that our approach substantially outperforms previous state-of-the-art techniques.

Keywords:
Generality Reinforcement learning Computer science Task (project management) Artificial intelligence Machine learning Robot Control (management) Scheme (mathematics) Action (physics) Policy learning Mathematics Engineering

Metrics

265
Cited By
0.00
FWCI (Field Weighted Citation Impact)
3
Refs
Citation Normalized Percentile
Is in top 1%
Is in top 10%

Citation History

Topics

Reinforcement Learning in Robotics
Physical Sciences →  Computer Science →  Artificial Intelligence
Adversarial Robustness in Machine Learning
Physical Sciences →  Computer Science →  Artificial Intelligence
Autonomous Vehicle Technology and Safety
Physical Sciences →  Engineering →  Automotive Engineering

Related Documents

JOURNAL ARTICLE

Data-Efficient Hierarchical Reinforcement Learning for Robotic Assembly Control Applications

Zhimin HouJiajun FeiYuelin DengJing Xu

Journal:   IEEE Transactions on Industrial Electronics Year: 2020 Vol: 68 (11)Pages: 11565-11575
JOURNAL ARTICLE

Towards Run-time Efficient Hierarchical Reinforcement Learning

Sasha AbramowitzGeoff Nitschke

Journal:   2022 IEEE Congress on Evolutionary Computation (CEC) Year: 2022 Vol: 97 Pages: 1-8
JOURNAL ARTICLE

Adjacency Constraint for Efficient Hierarchical Reinforcement Learning

Tianren ZhangShangqi GuoTian TanXiaolin HuFeng Chen

Journal:   IEEE Transactions on Pattern Analysis and Machine Intelligence Year: 2022 Vol: 45 (4)Pages: 4152-4166
JOURNAL ARTICLE

Data Efficient Safe Reinforcement Learning

Sindhu PadakandlaK. J. PrabuchandranS. GangulyShalabh Bhatnagar

Journal:   2022 IEEE International Conference on Systems, Man, and Cybernetics (SMC) Year: 2022 Vol: 16 Pages: 1167-1172
JOURNAL ARTICLE

Efficient Hierarchical Storage Management Empowered by Reinforcement Learning

Tianru ZhangAndreas HellanderSalman Toor

Journal:   IEEE Transactions on Knowledge and Data Engineering Year: 2022 Pages: 1-1
© 2026 ScienceGate Book Chapters — All rights reserved.