Kernel ε-Greedy for Multi-Armed Bandits with Covariates
Sakshi Arya, Bharath K. Sriperumbudur
Unverified — Be the first to reproduce this paper.
ReproduceAbstract
We consider the -greedy strategy for the multi-arm bandit with covariates (MABC) problem, where the mean reward functions are assumed to lie in a reproducing kernel Hilbert space (RKHS). We propose to estimate the unknown mean reward functions using an online weighted kernel ridge regression estimator, and show the resultant estimator to be consistent under appropriate decay rates of the exploration probability sequence, \_t\_t, and regularization parameter, \_t\_t. Moreover, we show that for any choice of kernel and the corresponding RKHS, we achieve a sub-linear regret rate depending on the intrinsic dimensionality of the RKHS. Furthermore, we achieve the optimal regret rate of T under a margin condition for finite-dimensional RKHS.