0


基于层次模型的强化学习

Hierarchical Model-Based Reinforcement Learning
课程网址: http://videolectures.net/icml08_jong_hmb/  
主讲教师: Nicholas Jong
开课单位: 德克萨斯大学
开课时间: 2008-08-04
课程语种: 英语
中文简介:
层次分解承诺通过利用底层结构,帮助将强化学习算法自然地扩展到现实问题。基于模型的算法为增强学习提供了第一个有限时间收敛的保证,在处理大环境中数据相对稀缺的问题上也可能发挥重要作用。本文介绍了一种在标准强化学习设置中充分集成现代层次学习和模型学习方法的算法。我们的算法R-MAXQ继承了基于模型的R-MAX算法的有效探索和MAXQ框架提供的抽象机会。我们分析了算法的样本复杂性,并在标准模拟环境中进行的实验说明了层次和模型相结合的优点。
课程简介: Hierarchical decomposition promises to help scale reinforcement learning algorithms naturally to real-world problems by exploiting their underlying structure. Model-based algorithms, which provided the first finite-time convergence guarantees for reinforcement learning, may also play an important role in coping with the relative scarcity of data in large environments. In this paper, we introduce an algorithm that fully integrates modern hierarchical and model-learning methods in the standard reinforcement learning setting. Our algorithm, R-maxq, inherits the efficient model-based exploration of the R-max algorithm and the opportunities for abstraction provided by the MAXQ framework. We analyze the sample complexity of our algorithm, and our experiments in a standard simulation environment illustrate the advantages of combining hierarchies and models.
关 键 词: 相对稀缺性; 强化学习; 学习框架
课程来源: 视频讲座网
最后编审: 2020-06-08:yumf
阅读次数: 76