期刊文献+

Extended Dyna-Q Algorithm for Path Planning of Mobile Robots

Extended Dyna-Q Algorithm for Path Planning of Mobile Robots
在线阅读 下载PDF
导出
摘要 This paper presents an extended Dyna-Q algorithm to improve efficiency of the standard Dyna-Q algorithm.In the first episodes of the standard Dyna-Q algorithm,the agent travels blindly to find a goal position.To overcome this weakness,our approach is to use a maximum likelihood model of all state-action pairs to choose actions and update Q-values in the first few episodes.Our algorithm is compared with one-step Q-learning algorithm and the standard Dyna-Q algorithm for the path planning problem in maze environments.Experimental results show that the proposed algorithm is more efficient than the one-step Q-learning algorithm as well as the standard Dyna-Q algorithm,especially in the large environment of states.
出处 《Journal of Measurement Science and Instrumentation》 CAS 2011年第3期283-287,共5页 测试科学与仪器(英文版)
基金 supported by the Basic Science Research Program through the National Research Foundation of Korea(NRF)funded by the Ministry of Education,Science and Technology(2010-0012609)
关键词 reinforcement learning Dyna-Q path planning mobile robots 机器人路径规划 Q算法 移动 学习算法 目标位置 最大似然 规划问题 标准
  • 相关文献

参考文献6

  • 1Sutton RS,Barto AG.Reinforcement Learning: An Introduction[]..1998
  • 2Smart W D,Kaelbling L P.Effective reinforcement learning for mobile robots[].Proceeding of the IEEE Intl Conference on robotics and Automation.2002
  • 3Fujisawa S,Kurozumi R,Yamamoto T,et al.Path planning for mobile robots using an improved reinforcement learning scheme[].Proceedings of the IEEE International Symposium on Intelligent Control.2002
  • 4Andrew W. Moore,Christopher G. Atkeson.Prioritized Sweeping: Reinforcement Learning with Less Data and Less Time[J].Machine Learning.1993(1)
  • 5C.Watkins.Learning from delayed rewards[]..1989
  • 6Kaelbling LP,Littman ML,Moore AW.Reinforcement learning: a survey[].Journal of Artificial Organs.1996

相关作者

内容加载中请稍等...

相关机构

内容加载中请稍等...

相关主题

内容加载中请稍等...

浏览历史

内容加载中请稍等...
;
使用帮助 返回顶部