首页    期刊浏览 2025年05月31日 星期六
登录注册

文章基本信息

  • 标题:Practical Reinforcement Learning -Experiences in Lot Scheduling Application ⁎
  • 本地全文:下载
  • 作者:Hannu Rummukainen ; Jukka K. Nurminen
  • 期刊名称:IFAC PapersOnLine
  • 印刷版ISSN:2405-8963
  • 出版年度:2019
  • 卷号:52
  • 期号:13
  • 页码:1415-1420
  • DOI:10.1016/j.ifacol.2019.11.397
  • 语种:English
  • 出版社:Elsevier
  • 摘要:With recent advances in deep reinforcement learning, it is time to take another look at reinforcement learning as an approach for discrete production control. We applied proximal policy optimization (PPO), a recently developed algorithm for deep reinforcement learning, to the stochastic economic lot scheduling problem. The problem involves scheduling manufacturing decisions on a single machine under stochastic demand, and despite its simplicity remains computationally challenging. We implemented two parameterized models for the control policy and value approximation, a linear model and a neural network, and used a modified PPO algorithm to seek the optimal parameter values. Benchmarking against the best known control policy for the test case, in which Paternina-Arboleda and Das (2005) combined a base-stock policy and an older reinforcement learning algorithm, we improved the average cost rate by 2 %. Our approach is more general, as we do not require a priori policy parameters such as base-stock levels, and the entire policy is learned.
  • 关键词:KeywordsReinforcement learningStochastic economic lot schedulingLearning controlStochastic controlMonte Carlo simulationNeural networksMachine learning
国家哲学社会科学文献中心版权所有