首页 | 本学科首页   官方微博 | 高级检索  
     


A note on strategies for bandit problems with infinitely many arms
Authors:Kung-Yu Chen  Chien-Tai Lin
Affiliation:(1) Department of Mathematics, Tamkang University, Tamsui, 251, Taiwan
Abstract:
A bandit problem consisting of a sequence of n choices (nrarrinfin) from a number of infinitely many Bernoulli arms is considered. The parameters of Bernoulli arms are independent and identically distributed random variables from a common distribution F on the interval [0,1] and F is continuous with F(0)=0 and F(1)=1. The goal is to investigate the asymptotic expected failure rates of k-failure strategies, and obtain a lower bound for the expected failure proportion over all strategies presented in Berry et al. (1997). We show that the asymptotic expected failure rates of k-failure strategies when 0<ble1 and a lower bound can be evaluated if the limit of the ratio F(1)–F(t) versus (1–t)b exists as trarr1 for some b>0.
Keywords:k-failure strategy  m-run strategy  Nn-learning strategy  non-recalling m-run strategy
本文献已被 SpringerLink 等数据库收录!
设为首页 | 免责声明 | 关于勤云 | 加入收藏

Copyright©北京勤云科技发展有限公司  京ICP备09084417号