This study reviews the major developments of Deep Deterministic Policy Gradient (DDPG) in the field of reinforcement learning. It is innovated by Deep Q-network ideas and can finally handle some much challenging problems that operate over continuous action space. The main idea of DDPG is to use an actor-critic architecture (shown in Figure 5) to learn much more competitive policies. It allows the model to use neural network function approximators to learn in large state and action spaces. Due to its strong capacity, DDPG has many useful applications to real world problems in the field like robotics and control systems. But like most of the model-free reinforcement learning methods, the requirement for a large number of training steps is still a major difficulty for DDPG.
Teckchai TiongIsmail SaadKenneth Tze Kin TeoHerwansyah bin Lago
Ming ZhanJingjing FanJianying Guo
Miao ChenWenna LiShihan FeiYufei WeiMingyang TuJiangbo Li
Shihui LiYi WuXinyue CuiHonghua DongFei FangStuart Russell