BOOK-CHAPTER

Gradient Correction for Asynchronous Stochastic Gradient Descent in Reinforcement Learning

Abstract

Abstract Distributed stochastic gradient descent techniques have gained significant attention in recent years as a prevalent approach for reinforcement learning. Current distributed learning predominantly employs synchronous or asynchronous training strategies. While the asynchronous scheme avoids idle computing resources present in synchronous methods, it grapples with the stale gradient issue. This paper introduces a novel gradient correction algorithm aimed at alleviating the stale gradient problem. By leveraging second-order information within the worker node and incorporating current parameters from both the worker and server nodes, the gradient correction algorithm yields a refined gradient closer to the desired value. Initially, we outline the challenges associated with asynchronous update schemes and derive a gradient correction algorithm employing local second-order approximations. Subsequently, we propose an asynchronous training scheme incorporating gradient correction within the generalized policy iteration framework. Lastly, in the context of trajectory tracking tasks, we compare the impact of employing gradient correction versus its absence in an asynchronous update scheme. Simulation results underscore the superiority of our proposed training scheme, demonstrating notably faster convergence and higher policy performance compared to the existing asynchronous update methods.

Keywords:
Reinforcement learning Asynchronous communication Gradient descent Stochastic gradient descent Asynchronous learning Computer science Descent (aeronautics) Artificial intelligence Mathematics Physics Artificial neural network Mathematics education Telecommunications Meteorology

Metrics

0
Cited By
0.00
FWCI (Field Weighted Citation Impact)
10
Refs
0.38
Citation Normalized Percentile
Is in top 1%
Is in top 10%

Topics

Reinforcement Learning in Robotics
Physical Sciences →  Computer Science →  Artificial Intelligence
Advanced Memory and Neural Computing
Physical Sciences →  Engineering →  Electrical and Electronic Engineering
Neuroscience and Neural Engineering
Life Sciences →  Neuroscience →  Cellular and Molecular Neuroscience

Related Documents

JOURNAL ARTICLE

Asynchronous Decentralized Accelerated Stochastic Gradient Descent

Guanghui LanYi Zhou

Journal:   IEEE Journal on Selected Areas in Information Theory Year: 2021 Vol: 2 (2)Pages: 802-811
JOURNAL ARTICLE

Asynchronous Stochastic Gradient Descent Over Decentralized Datasets

Yubo DuKeyou You

Journal:   IEEE Transactions on Control of Network Systems Year: 2021 Vol: 8 (3)Pages: 1212-1224
© 2026 ScienceGate Book Chapters — All rights reserved.