Explicit Dropout: Deterministic Regularization for Transformer Architectures
arXiv:2604.20505v1 Announce Type: new
Abstract: Dropout is a widely used regularization technique in deep learning, but its effects are typically realized through stochastic masking rather than explicit optimization objectives. We propose a determinis…