首页    期刊浏览 2024年09月15日 星期日
登录注册

文章基本信息

  • 标题:A reinforcement learning method with closed-loop stability guarantee for systems with unknown parameters
  • 本地全文:下载
  • 作者:Thomas Göhrt ; Fritjof Griesing-Scheiwe ; Pavel Osinenko
  • 期刊名称:IFAC PapersOnLine
  • 印刷版ISSN:2405-8963
  • 出版年度:2020
  • 卷号:53
  • 期号:2
  • 页码:8157-8162
  • DOI:10.1016/j.ifacol.2020.12.2303
  • 语种:English
  • 出版社:Elsevier
  • 摘要:AbstractThis work is concerned with the application of reinforcement learning (RL) techniques to adaptive dynamic programming (ADP) for systems with partly unknown models. In ADP, one seeks to approximate an optimal infinite horizon cost function, the value function. Such an approximation, i.e., critic, does not in general yield a stabilizing control policies, i.e., stabilizing actors. Guaranteeing stability of nonlinear systems under RL/ADP is still an open issue. In this work, it is suggested to use a stability constraint directly in the actor-critic structure. The system model considered in this work is assumed to be only partially known, specifically, it contains an unknown parameter vector. A suitable stabilizability assumption for such systems is an adaptive Lyapunov function, which is commonly assumed in adaptive control. The current approach formulates a stability constraint based on an adaptive Lyapunov function to ensure closed-loop stability. Convergence of the actor and critic parameters in a suitable sense is shown. A case study demonstrates how the suggested algorithm preserves closed-loop stability, while at the same time improving an infinite-horizon performance.
  • 关键词:KeywordsConsensusReinforcement learning controlNonlinear adaptive control
国家哲学社会科学文献中心版权所有