cs.LG

MARBLE: Multi-Armed Restless Bandits in Latent Markovian Environment

arXiv:2511.09324v2 Announce Type: replace
Abstract: Restless Multi-Armed Bandits (RMABs) are powerful models for decision-making under uncertainty, yet classical formulations typically assume fixed dynamics, an assumption often violated in nonstationa…