通信学报 ›› 2016, Vol. 37 ›› Issue (6): 65-74.doi: 10.11959/j.issn.1000-436x.2016117

• 学术论文 • 上一篇    下一篇

基于可中断Option的在线分层强化学习方法

朱斐1,2,许志鹏1,刘全1,2,伏玉琛1,王辉1   

  1. 1 苏州大学计算机科学与技术学院,江苏 苏州215006
    2 吉林大学符号计算与知识工程教育部重点实验室,吉林 长春130012
  • 出版日期:2016-06-25 发布日期:2017-08-04
  • 基金资助:
    国家自然科学基金资助项目;国家自然科学基金资助项目;国家自然科学基金资助项目;国家自然科学基金资助项目;江苏省高校自然科学研究基金资助项目;吉林大学符号计算与知识工程教育部重点实验室基金资助项目;苏州市应用基础研究计划基金资助项目;苏州大学高校省级重点实验室基金资助项目;中国国家留学基金资助项目

Online hierarchical reinforcement learning based on interrupting Option

Fei ZHU1,2,Zhi-peng XU1,Quan LIU1,2,Yu-chen FU1,Hui WANG1   

  1. 1 School of Computer Science and Technology,Soochow University,Suzhou 215006,China
    2 Key Laboratory of Symbolic Computation and Knowledge Engineering of Ministry of Education,Jilin University,Changchun 130012,China
  • Online:2016-06-25 Published:2017-08-04
  • Supported by:
    The National Natural Science Foundation of China;The National Natural Science Foundation of China;The National Natural Science Foundation of China;The National Natural Science Foundation of China;The High School Natural Foundation of Jiangsu Province;The Key Laboratory of Symbolic Computation and Knowledge Engineering of Ministry of Education of Jilin University;Suzhou Industrial Application of Basic Research Program;Provincial Key Laboratory for Computer Information Processing Technology of Soochow University;The China Scholarship Council Project

摘要:

针对大数据体量大的问题,在Macro-Q算法的基础上提出了一种在线更新的Macro-Q算法(MQIU),同时更新抽象动作的值函数和元动作的值函数,提高了数据样本的利用率。针对传统的马尔可夫过程模型和抽象动作均难于应对可变性,引入中断机制,提出了一种可中断抽象动作的Macro-Q无模型学习算法(IMQ),能在动态环境下学习并改进控制策略。仿真结果验证了MQIU算法能加快算法收敛速度,进而能解决更大规模的问题,同时也验证了IMQ算法能够加快任务的求解,并保持学习性能的稳定性。

关键词: 大数据, 强化学习, 分层强化学习, Option, 在线学习

Abstract:

Aiming at dealing with volume of big data,an on-line updating algorithm,named by Macro-Q with in-place updating (MQIU),which was based on Macro-Q algorithm and takes advantage of in-place updating approach,was proposed.The MQIU algorithm updates both the value function of abstract action and the value function of primitive action,and hence speeds up the convergence rate.By introducing the interruption mechanism,a model-free interrupting Macro-Q Option learning algorithm(IMQ),which was based on hierarchical reinforcement learning,was also introduced to order to handle the variability which was hard to process by the conventional Markov decision process model and abstract action so that IMQ was able to learn and improve control strategies in a dynamic environment.Simulations verify the MQIU algorithm speeds up the convergence rate so that it is able to do with the larger scale of data,and the IMQ algorithm solves the task faster with a stable learning performance.

Key words: big data, reinforcement learning, hierarchical reinforcement learning, Option, online learning

No Suggested Reading articles found!