Off-Policy Risk-Sensitive Reinforcement Learning-Based Constrained Robust Optimal Control

Cong Li, Qingchen Liu, Zhehua Zhou, Martin Buss, Fangzhou Liu

Research output: Contribution to journalArticlepeer-review

2 Scopus citations

Abstract

This article proposes an off-policy risk-sensitive reinforcement learning (RL)-based control framework to jointly optimize the task performance and constraint satisfaction in a disturbed environment. The risk-aware value function, constructed using the pseudo control and risk-sensitive input and state penalty terms, is introduced to convert the original constrained robust stabilization problem into an equivalent unconstrained optimal control problem. Then, an off-policy RL algorithm is developed to learn the approximate solution to the risk-aware value function. During the learning process, the associated approximate optimal control policy is able to satisfy both input and state constraints under disturbances. By replaying experience data to the off-policy weight update law of the critic neural network, the weight convergence is guaranteed. Moreover, online and offline algorithms are developed to serve as principled ways to record informative experience data to achieve a sufficient excitation required for the weight convergence. The proofs of system stability and weight convergence are provided. The Simulation results reveal the validity of the proposed control framework.

Original languageEnglish
Pages (from-to)2478-2491
Number of pages14
JournalIEEE Transactions on Systems, Man, and Cybernetics: Systems
Volume53
Issue number4
DOIs
StatePublished - 1 Apr 2023

Keywords

  • Adaptive dynamic programming (ADP)
  • input saturation
  • off-policy risk-sensitive reinforcement learning (RL)
  • robust control
  • state constraint

Fingerprint

Dive into the research topics of 'Off-Policy Risk-Sensitive Reinforcement Learning-Based Constrained Robust Optimal Control'. Together they form a unique fingerprint.

Cite this