通信学报 ›› 2013, Vol. 34 ›› Issue (10): 92-99.doi: 10.3969/j.issn.1000-436x.2013.10.011

• 学术论文 • 上一篇    下一篇

基于两层模糊划分的时间差分算法

穆翔1,刘全1,2,傅启明1,孙洪坤1,周鑫1   

  1. 1 苏州大学 计算机科学与技术学院,江苏 苏州 215006
    2 吉林大学 符号计算与知识工程教育部重点实验室,吉林 长春 130012
  • 出版日期:2013-10-25 发布日期:2017-08-10
  • 基金资助:
    国家自然科学基金资助项目;国家自然科学基金资助项目;国家自然科学基金资助项目;国家自然科学基金资助项目;江苏省自然科学基金资助项目;江苏省高校自然科学研究基金资助项目;江苏省高校自然科学研究基金资助项目;吉林大学符号计算与知识工程教育部重点实验室基金资助项目

TD algorithm based on double-layer fuzzy partitioning

Xiang MU1,Quan LIU1,2,Qi-ming FU1,Hong-kun SUN1,Xin ZHOU1   

  1. 1 Institute of Computer Science and Technology,Soochow University,Suzhou 215006,China
    2 Key Laboratory of Symbolic Computation and Knowledge Engineering of Ministry of Education,Jilin University,Changchun 130012,China
  • Online:2013-10-25 Published:2017-08-10
  • Supported by:
    The National Natural Science Foundation of China;The National Natural Science Foundation of China;The National Natural Science Foundation of China;The National Natural Science Foundation of China;The Natural Science Foundation of Jiangsu Province;The High School Natural Foundation of Jiangsu Province;The High School Natural Foundation of Jiangsu Province;The Foundation of Key Laboratory of Symbolic Computation and Knowledge Engineering of Ministry of Education of Jilin University

摘要:

针对传统的基于查询表或函数逼近的Q值迭代算法在处理连续空间问题时收敛速度慢、且不易求解连续行为策略的问题,提出了一种基于两层模糊划分的在策略时间差分算法——DFP-OPTD,并从理论上分析其收敛性。算法中第一层模糊划分作用于状态空间,第二层模糊划分作用于动作空间,并结合两层模糊划分计算出Q值函数。根据所得的 值函数,使用梯度下降方法更新模糊规则中的后件参数。将Q DFP-OPTD应用于经典强化学习问题中,实验结果表明,该算法有较好的收敛性能,且可以求解连续行为策略。

关键词: 强化学习, 在策略, 梯度下降, 两层模糊划分, 连续行为策略

Abstract:

When dealing with the continuous space problems,the traditional Q-iteration algorithms based on lookup-table or function approximation converge slowly and are diff lt to get a continuous policy.To overcome the above weak-nesses,an on-policy TD algorithm named DFP-OPTD was proposed based on double-layer fuzzy partitioning and its convergence was proved.The first layer of fuzzy partitioning was applied for state space,the second layer of fuzzy parti-tioning was applied for action space,and Q-value functions were computed by the combination of the two layer fuzzy partitioning.Based on the Q-value function,the consequent parameters of fuzzy rules were updated by gradient descent method.Applying DFP-OPTD on two classical reinforcement learning problems,experimental results show that the algo-rithm not only can be used to get a continuous action policy,but also has a better convergence performance.

Key words: reinforcement learning, on-policy, gradient descent;, double layer fuzzy partitioning, continuous action policy

No Suggested Reading articles found!