高级检索
    章晓芳, 周倩, 梁斌, 徐进. 一种自适应的多臂赌博机算法[J]. 计算机研究与发展, 2019, 56(3): 643-654. DOI: 10.7544/issn1000-1239.2019.20180019
    引用本文: 章晓芳, 周倩, 梁斌, 徐进. 一种自适应的多臂赌博机算法[J]. 计算机研究与发展, 2019, 56(3): 643-654. DOI: 10.7544/issn1000-1239.2019.20180019
    Zhang Xiaofang, Zhou Qian, Liang Bin, Xu Jin. An Adaptive Algorithm in Multi-Armed Bandit Problem[J]. Journal of Computer Research and Development, 2019, 56(3): 643-654. DOI: 10.7544/issn1000-1239.2019.20180019
    Citation: Zhang Xiaofang, Zhou Qian, Liang Bin, Xu Jin. An Adaptive Algorithm in Multi-Armed Bandit Problem[J]. Journal of Computer Research and Development, 2019, 56(3): 643-654. DOI: 10.7544/issn1000-1239.2019.20180019

    一种自适应的多臂赌博机算法

    An Adaptive Algorithm in Multi-Armed Bandit Problem

    • 摘要: 多臂赌博机问题是强化学习中研究探索和利用两者平衡的经典问题,其中,随机多臂赌博机问题是最经典的一类多臂赌博机问题,是众多新型多臂赌博机问题的基础.针对现有多臂赌博机算法未能充分使用环境反馈信息以及泛化能力较弱的问题,提出一种自适应的多臂赌博机算法.该算法利用当前估计值最小的动作被选择的次数来调整探索和利用的概率(chosen number of arm with minimal estimation, CNAME),有效缓解了探索和利用不平衡的问题.同时,该算法不依赖于上下文信息,在不同场景的多臂赌博机问题中有更好的泛化能力.通过理论分析给出了该算法的悔值(regret)上界,并通过不同场景的实验结果表明:CNAME算法可以高效地获得较高的奖赏和较低的悔值,并且具有更好的泛化能力.

       

      Abstract: As an important ongoing field in machine learning, reinforcement learning has received extensive attention in recent years. The multi-armed bandit (MAB) problem is a typical problem of the exploration and exploitation dilemma in reinforcement learning. As a classical MAB problem, the stochastic multi-armed bandit (SMAB) problem is the base of many new MAB problems. To solve the problems of insufficient use of information and poor generalization ability in existing MAB methods, this paper presents an adaptive SMAB algorithm to balance exploration and exploitation based on the chosen number of arm with minimal estimation, namely CNAME in short. CNAME makes use of the chosen times and the estimations of an action at the same time, so that an action is chosen according to the exploration probability, which is updated adaptively. In order to control the decline rate of exploration probability, the parameter w is introduced to adjust the influence degree of feedback during the selection process. Furthermore, CNAME does not depend on contextual information, hence it has better generalization ability. The upper bound of CNAMEs regret is theoretically proved and analyzed. Our experimental results in different scenarios show that CNAME can yield greater reward and smaller regret with high efficiency than commonly used methods. In addition, its generalization ability is very strong.

       

    /

    返回文章
    返回