Sessa: Selective State Space Attention
arXiv:2604.18580v2 Announce Type: replace
Abstract: Modern sequence modeling is dominated by two families: Transformers, whose self-attention can access arbitrary elements of the visible sequence, and structured state-space models, which propagate inf…