[edit]
When and why randomised exploration works (in linear bandits)
Proceedings of The 36th International Conference on Algorithmic Learning Theory, PMLR 272:4-22, 2025.
Abstract
We provide an approach for the analysis of randomised exploration algorithms like Thompson sampling that does not rely on forced optimism or posterior inflation. With this, we demonstrate that in the d-dimensional linear bandit setting, when the action space is smooth and strongly convex, randomised exploration algorithms enjoy an n-step regret bound of the order O(d√nlog(n)). Notably, this shows for the first time that there exist non-trivial linear bandit settings where Thompson sampling can achieve optimal dimension dependence in the regret.