JOURNAL ARTICLE

Multi-goal Reinforcement Learning via Exploring Successor Matching

Abstract

Multi-goal reinforcement learning (RL) agent aims at achieving and generalizing over various goals. Due to the sparsity of goal-reaching rewards, it suffers from unreliable value estimation and is thus unable to efficiently identify essential states towards specific goal-reaching. To deal with the problem, we propose Exploring Successor Matching (ESM), a framework that enables goal-conditioned policy and progressively encourages the multi-goal exploration towards the promising frontier. ESM adopts the idea of successor feature and extends it to goal-reaching successor mapping that serves as a more stable state feature under sparse rewards. After acquiring the successor mapping, it further explores intrinsic goals that are more likely to be achieved from a diverse set of states in terms of future state occupancies. Experiments on challenging manipulation tasks show that ESM deals well with sparse rewards and achieves better sample efficiency.

Keywords:
Successor cardinal Reinforcement learning Computer science Matching (statistics) Set (abstract data type) Artificial intelligence Reinforcement Machine learning Feature (linguistics) Goal orientation State (computer science) Mathematics Psychology Algorithm Social psychology

Metrics

2
Cited By
0.39
FWCI (Field Weighted Citation Impact)
92
Refs
0.61
Citation Normalized Percentile
Is in top 1%
Is in top 10%

Citation History

Topics

Reinforcement Learning in Robotics
Physical Sciences →  Computer Science →  Artificial Intelligence
Robot Manipulation and Learning
Physical Sciences →  Engineering →  Control and Systems Engineering
Adaptive Dynamic Programming Control
Physical Sciences →  Computer Science →  Computational Theory and Mathematics

Related Documents

© 2026 ScienceGate Book Chapters — All rights reserved.