LoopVLA: Learning Sufficiency in Recurrent Refinement for Vision-Language-Action Models
arXiv:2605.09948v1 Announce Type: new
Abstract: Current Vision-Language-Action (VLA) models typically treat the deepest representation of a vision-language backbone as universally optimal for action prediction. However, robotic manipulation is compose…