JOURNAL ARTICLE

Curiosity-driven exploration in reinforcement learning

Abstract

The paper elaborates upon a prior proposal for a novelty detector based on an artificial neural network forecaster. In the former paper, the novelty-based motivational signal was used in place of more conventional techniques (such as the ε-greedy policy, or the softmax policy) to drive exploration, in the context of V-learning. The current paper provides a more comprehensive study of such handling of the exploration vs. exploitation trade-off. It also studies the various problems concerning application of the approach to SARSA, and Q-learning. Also, and with the same goal in mind, the paper presents several advances upon the original design.

Keywords:
Novelty Reinforcement learning Curiosity Softmax function Computer science Artificial intelligence Context (archaeology) Novelty detection Machine learning Artificial neural network Psychology

Metrics

7
Cited By
1.13
FWCI (Field Weighted Citation Impact)
13
Refs
0.83
Citation Normalized Percentile
Is in top 1%
Is in top 10%

Citation History

Topics

Advanced Bandit Algorithms Research
Social Sciences →  Decision Sciences →  Management Science and Operations Research
Fault Detection and Control Systems
Physical Sciences →  Engineering →  Control and Systems Engineering
Reinforcement Learning in Robotics
Physical Sciences →  Computer Science →  Artificial Intelligence
© 2026 ScienceGate Book Chapters — All rights reserved.