Latent Bandits
暂无分享,去创建一个
[1] Jean-Yves Audibert,et al. Deviations of Stochastic Bandit Regret , 2011, ALT.
[2] Eric Moulines,et al. A Near Optimal Policy for Channel Allocation in Cognitive Radio , 2008, EWRL.
[3] D. Teneketzis,et al. Asymptotically Efficient Adaptive Allocation Schemes for Controlled I.I.D. Processes: Finite Paramet , 1988 .
[4] Csaba Szepesvári,et al. Improved Algorithms for Linear Stochastic Bandits , 2011, NIPS.
[5] Shie Mannor,et al. Decoupling Exploration and Exploitation in Multi-Armed Bandits , 2012, ICML.
[6] R. Munos,et al. Kullback–Leibler upper confidence bounds for optimal sequential allocation , 2012, 1210.1136.
[7] T. L. Lai Andherbertrobbins. Asymptotically Efficient Adaptive Allocation Rules , 2022 .
[8] Martin Pál,et al. Contextual Multi-Armed Bandits , 2010, AISTATS.
[9] Thomas P. Hayes,et al. Stochastic Linear Optimization under Bandit Feedback , 2008, COLT.
[10] A. Burnetas,et al. Optimal Adaptive Policies for Sequential Allocation Problems , 1996 .
[11] J. Langford,et al. The Epoch-Greedy algorithm for contextual multi-armed bandits , 2007, NIPS 2007.
[12] Gediminas Adomavicius,et al. Toward the next generation of recommender systems: a survey of the state-of-the-art and possible extensions , 2005, IEEE Transactions on Knowledge and Data Engineering.
[13] Wei Chu,et al. A contextual-bandit approach to personalized news article recommendation , 2010, WWW '10.
[14] Aleksandrs Slivkins,et al. Contextual Bandits with Similarity Information , 2009, COLT.
[15] Wei Chu,et al. Unbiased offline evaluation of contextual-bandit-based news article recommendation algorithms , 2010, WSDM '11.
[16] Peter Auer,et al. Using Confidence Bounds for Exploitation-Exploration Trade-offs , 2003, J. Mach. Learn. Res..
[17] Nimrod Megiddo,et al. Online Learning with Prior Knowledge , 2007, COLT.
[18] John Langford,et al. The Epoch-Greedy Algorithm for Multi-armed Bandits with Side Information , 2007, NIPS.
[19] Aurélien Garivier,et al. On Upper-Confidence Bound Policies for Non-Stationary Bandit Problems , 2008 .