Actor–critic learning based PID control for robotic manipulators

Nohooji, Hamed Rahimi, Zaraki, Abolfazl and Voos, Holger (2024) Actor–critic learning based PID control for robotic manipulators. Applied Soft Computing, 151: 111153. pp. 1-11. ISSN 1568-4946
Copy

In this paper, we propose a reinforcement learning structure for auto-tuning PID gains by solving an optimal tracking control problem for robot manipulators. Capitalizing on the actor–critic framework implemented by neural networks, we achieve optimal tracking performance, estimating unknown system dynamics simultaneously. The critic network is used to approximate the cost function, which serves as an indicator of control performance. With feedback from the critic, the actor network learns time-varying PID gains over time to optimize control input, thereby steering the system toward optimal performance. Furthermore, we utilize Lyapunov's direct method to demonstrate the stability of the closed-loop system. This approach provides an analytical procedure for a stable robot manipulator system to systematically adjust PID gains, bypassing the ad-hoc and painstaking process. The resultant actor–critic PID-like control exhibits stable adaptive and learning capabilities while maintaining a simple structure and inexpensive online computational demands. Numerical simulations underscore the effectiveness and advantages of the proposed actor–critic neural network PID control.

visibility_off picture_as_pdf

picture_as_pdf
ASOC-D-23-01115_R3.pdf
subject
Submitted Version
lock_clock
Restricted to Repository staff only until 14 December 2025
Available under Creative Commons: BY-NC-ND 4.0

Request Copy

Atom BibTeX OpenURL ContextObject in Span OpenURL ContextObject Dublin Core MPEG-21 DIDL Data Cite XML EndNote HTML Citation METS MODS RIOXX2 XML Reference Manager Refer ASCII Citation
Export

Downloads