MARBLE: Multi-Armed Restless Bandits in Latent Markovian Environment
arXiv:2511.09324v2 Announce Type: replace
Abstract: Restless Multi-Armed Bandits (RMABs) are powerful models for decision-making under uncertainty, yet classical formulations typically assume fixed dynamics, an assumption often violated in nonstationa…