首页    期刊浏览 2024年10月05日 星期六
登录注册

文章基本信息

  • 标题:Reinforcement Learning for mixed-integer problems based on MPC
  • 本地全文:下载
  • 作者:Sebastien Gros ; Mario Zanon
  • 期刊名称:IFAC PapersOnLine
  • 印刷版ISSN:2405-8963
  • 出版年度:2020
  • 卷号:53
  • 期号:2
  • 页码:5219-5224
  • DOI:10.1016/j.ifacol.2020.12.1196
  • 语种:English
  • 出版社:Elsevier
  • 摘要:AbstractModel Predictive Control has been recently proposed as policy approximation for Reinforcement Learning, offering a path towards safe and explainable Reinforcement Learning. This approach has been investigated for Q-learning and actor-critic methods, both in the context of nominal economic MPC and Robust (N)MPC, showing very promising results. In that context, actor-critic methods seem to be the most reliable approach. Many applications include a mixture of continuous and integer inputs, for which the classical actor-critic methods need to be adapted. In this paper, we present a policy approximation based on mixed-integer MPC schemes, and propose a computationally inexpensive technique to generate exploration in the mixed-integer input space that ensures a satisfaction of the constraints. We then propose a simple compatible advantage function approximation for the proposed policy, that allows one to build the gradient of the mixed-integer MPC-based policy.
  • 关键词:KeywordsReinforcement LearningMixed-Integer Model Predictive Controlactor-critic methodsstochasticdeterministic policy gradient
国家哲学社会科学文献中心版权所有