Can Explicit Physical Feasibility Benefit VLA Learning? An Empirical Study
arXiv:2604.17896v1 Announce Type: cross
Abstract: Vision-Language-Action (VLA) models map multimodal inputs directly to robot actions and are typically trained through large-scale imitation learning. While this paradigm has shown strong performance, p…