JOURNAL ARTICLE

Sparse Bayesian Network-Based Disturbance Observer for Policy-Based Reinforcement Learning

Abstract

We proposed the Sparse Bayesian Network-based Disturbance Observer (SBN-DOB) to enhance the robustness of policy-based reinforcement learning. SBN-DOB utilizes sparse Bayesian learning to estimate the nominal inverse model dynamics, effectively mitigating model uncertainty and disturbances without relying on physical modeling. The SBN-DOB can be compressed by inducing sparsity in network parameters through sparse Bayesian learning, and the Bayesian model reduces the risk of overfitting during inference. To evaluate the effectiveness of the proposed approach, We conducted experiments using the policy network (PN) of the soft actor-critic algorithm in combination with SBN-DOB for six control tasks with an uncertain environment. The results of these experiments demonstrate that the performance of PN is preserved against continuous disturbances and state noise even with compression applied to SBN-DOB. Consequently, SBN-DOB is expected to minimize the simulation-to-reality gap of reinforcement learning by used in embedded systems with limited performance and capacity.

Keywords:
Reinforcement learning Computer science Overfitting Bayesian probability Artificial intelligence Bayesian inference Robustness (evolution) Dynamic Bayesian network Machine learning Q-learning Inference Posterior probability Artificial neural network

Metrics

0
Cited By
0.00
FWCI (Field Weighted Citation Impact)
10
Refs
0.15
Citation Normalized Percentile
Is in top 1%
Is in top 10%

Topics

Reinforcement Learning in Robotics
Physical Sciences →  Computer Science →  Artificial Intelligence
Adaptive Dynamic Programming Control
Physical Sciences →  Computer Science →  Computational Theory and Mathematics
Advanced Control Systems Optimization
Physical Sciences →  Engineering →  Control and Systems Engineering
© 2026 ScienceGate Book Chapters — All rights reserved.