Offline Local Search for Online Stochastic Bandits
arXiv:2604.09423v1 Announce Type: new
Abstract: Combinatorial multi-armed bandits provide a fundamental online decision-making environment where a decision-maker interacts with an environment across $T$ time steps, each time selecting an action and le…