Actor–critic learning based PID control for robotic manipulators
In this paper, we propose a reinforcement learning structure for auto-tuning PID gains by solving an optimal tracking control problem for robot manipulators. Capitalizing on the actor–critic framework implemented by neural networks, we achieve optimal tracking performance, estimating unknown system dynamics simultaneously. The critic network is used to approximate the cost function, which serves as an indicator of control performance. With feedback from the critic, the actor network learns time-varying PID gains over time to optimize control input, thereby steering the system toward optimal performance. Furthermore, we utilize Lyapunov's direct method to demonstrate the stability of the closed-loop system. This approach provides an analytical procedure for a stable robot manipulator system to systematically adjust PID gains, bypassing the ad-hoc and painstaking process. The resultant actor–critic PID-like control exhibits stable adaptive and learning capabilities while maintaining a simple structure and inexpensive online computational demands. Numerical simulations underscore the effectiveness and advantages of the proposed actor–critic neural network PID control.
Item Type | Article |
---|---|
Additional information | © 2023 Elsevier B.V. All rights reserved. This is the accepted manuscript version of an article which has been published in final form at https://doi.org/10.1016/j.asoc.2023.111153 |
Keywords | actor–critic, neural network, pid control, reinforcement learning, robot manipulators, software |
Date Deposited | 15 May 2025 15:52 |
Last Modified | 31 May 2025 00:47 |
-
picture_as_pdf - ASOC-D-23-01115_R3.pdf
-
subject - Submitted Version
-
lock_clock - Restricted to Repository staff only until 14 December 2025
-
- Available under Creative Commons: BY-NC-ND 4.0