首页    期刊浏览 2025年07月17日 星期四
登录注册

文章基本信息

  • 标题:A Temporal-Difference Learning Method Using Gaussian State Representation for Continuous State Space Problems
  • 本地全文:下载
  • 作者:Natsuko Fujii ; Atsushi Ueno ; Tomohito Takubo
  • 期刊名称:人工知能学会論文誌
  • 印刷版ISSN:1346-0714
  • 电子版ISSN:1346-8030
  • 出版年度:2014
  • 卷号:29
  • 期号:1
  • 页码:157-167
  • DOI:10.1527/tjsai.29.157
  • 出版社:The Japanese Society for Artificial Intelligence
  • 摘要:In this paper, we tackle the problem of reinforcement learning (RL) in a continuous state space. An appropriate discretization of the space can make many learning tasks tractable. A method using Gaussian state representation and the Rational Policy Making algorithm (RPM) has been proposed for this problem. This method discretizes the space by constructing a chain of states which represents a path to the goal of the agent exploiting past experiences of reaching it. This method exploits successful experiences strongly. Therefore, it can find a rational solution quickly in an environment with few noises. In a noisy environment, it makes many unnecessary and distractive states and does the task poorly. For learning in such an environment, we have introduced the concept of the value of a state to the above method and developed a new method. This method uses a temporal-difference (TD) learning algorithm for learning the values of states. The value of a state is used to determine the size of the state. Thus, our developed method can trim and eliminate unnecessary and distractive states quickly and learn the task well even in a noisy environment. We show the effectiveness of our method by computer simulations of a path finding task and a cart-pole swing-up task.
  • 关键词:reinforcement learning ; continuous state spaces ; gaussian state representation ; TD learning
国家哲学社会科学文献中心版权所有