BOOK-CHAPTER

Hierarchical Reinforcement Learning

Abstract

Reinforcement learning (RL) deals with the problem of an agent that has to learn how to behave to maximize its utility by its interactions with an environment (Sutton & Barto, 1998; Kaelbling, Littman & Moore, 1996). Reinforcement learning problems are usually formalized as Markov Decision Processes (MDP), which consist of a finite set of states and a finite number of possible actions that the agent can perform. At any given point in time, the agent is in a certain state and picks an action. It can then observe the new state this action leads to, and receives a reward signal. The goal of the agent is to maximize its long-term reward. In this standard formalization, no particular structure or relationship between states is assumed. However, learning in environments with extremely large state spaces is infeasible without some form of generalization. Exploiting the underlying structure of a problem can effect generalization and has long been recognized as an important aspect in representing sequential decision tasks (Boutilier et al., 1999). Hierarchical Reinforcement Learning is the subfield of RL that deals with the discovery and/or exploitation of this underlying structure. Two main ideas come into play in hierarchical RL. The first one is to break a task into a hierarchy of smaller subtasks, each of which can be learned faster and easier than the whole problem. Subtasks can also be performed multiple times in the course of achieving the larger task, reusing accumulated knowledge and skills. The second idea is to use state abstraction within subtasks: not every task needs to be concerned with every aspect of the state space, so some states can actually be abstracted away and treated as the same for the purpose of the given subtask. Request access from your librarian to read this chapter's full text.

Keywords:
Reinforcement learning Generalization Hierarchy Computer science Markov decision process Task (project management) Set (abstract data type) Artificial intelligence Action (physics) State (computer science) Abstraction Machine learning Markov process Mathematics Algorithm Engineering

Metrics

3
Cited By
0.00
FWCI (Field Weighted Citation Impact)
0
Refs
0.39
Citation Normalized Percentile
Is in top 1%
Is in top 10%

Citation History

Topics

Supply Chain and Inventory Management
Social Sciences →  Business, Management and Accounting →  Management Information Systems
Complex Systems and Decision Making
Social Sciences →  Decision Sciences →  Management Science and Operations Research

Related Documents

BOOK-CHAPTER

Hierarchical Reinforcement Learning

Bernhard Hengst

Encyclopedia of Machine Learning and Data Mining Year: 2017 Pages: 611-619
BOOK-CHAPTER

Hierarchical Reinforcement Learning

Carlos DiukMichael L. Littman

IGI Global eBooks Year: 2009 Pages: 825-830
BOOK-CHAPTER

Hierarchical Reinforcement Learning

Bernhard Hengst

Encyclopedia of Machine Learning Year: 2010 Pages: 495-502
BOOK-CHAPTER

Hierarchical Reinforcement Learning

Aske Plaat

Year: 2022 Pages: 263-285
BOOK-CHAPTER

Hierarchical Reinforcement Learning

Kerstin UngerDavid Badre

Elsevier eBooks Year: 2015 Pages: 367-373
© 2026 ScienceGate Book Chapters — All rights reserved.