cs.LG

Explicit Dropout: Deterministic Regularization for Transformer Architectures

arXiv:2604.20505v1 Announce Type: new
Abstract: Dropout is a widely used regularization technique in deep learning, but its effects are typically realized through stochastic masking rather than explicit optimization objectives. We propose a determinis…