首页    期刊浏览 2025年02月22日 星期六
登录注册

文章基本信息

  • 标题:Deep reinforcement learning based finite-horizon optimal tracking control for nonlinear system
  • 本地全文:下载
  • 作者:Jong Woo Kim ; Byung Jun Park ; Haeun Yoo
  • 期刊名称:IFAC PapersOnLine
  • 印刷版ISSN:2405-8963
  • 出版年度:2018
  • 卷号:51
  • 期号:25
  • 页码:257-262
  • DOI:10.1016/j.ifacol.2018.11.115
  • 语种:English
  • 出版社:Elsevier
  • 摘要:AbstractReinforcement learning (RL) can be used to obtain an approximate numerical solution to the Hamilton-Jacobi-Bellman (HJB) equation. Recent advances in machine learning community enable the use of deep neural networks (DNNs) to approximate high-dimensional nonlinear functions as those that occur in RL, accurately without any domain knowledge. In the standard RL setting, both system and cost structures are unknown, and the amount of data needed to obtain an accurate approximation can be impractically large. Meanwhile, when the structures are known, they can be used to solve the HJB equation efficiently. Herein, the model-based globalized dual heuristic programming (GDHP) is proposed, in which the HJB equation is separated into value, costate, and policy functions. A particular class of interest in this research is finite horizon optimal tracking control (FHOC) problem. Additional issues that arise, such as time-varying functions, terminal constraints, and delta-input formulation, are addressed in the context of FHOC. The DNN structure and training algorithm suitable for FHOC are presented. A benchmark continuous reactor example is provided to illustrate the proposed approach.
  • 关键词:KeywordsReinforcement learningApproximate dynamic programmingDeep learningGlobalized dual heuristic programmingOptimal controlOptimal tracking
国家哲学社会科学文献中心版权所有