SOTAVerified

Hard Attention

Papers

Showing 6170 of 100 papers

TitleStatusHype
NoPE: The Counting Power of Transformers with No Positional Encodings0
Achieving Explainability in a Visual Hard Attention Model through Content Prediction0
A Differentiable Self-disambiguated Sense Embedding Model via Scaled Gumbel Softmax0
AMR Parsing with Action-Pointer Transformer0
An Exploration of Neural Sequence-to-Sequence Architectures for Automatic Post-Editing0
A study of latent monotonic attention variants0
AttentionDrop: A Novel Regularization Method for Transformer Models0
Average-Hard Attention Transformers are Constant-Depth Uniform Threshold Circuits0
Characterizing the Expressivity of Transformer Language Models0
CLAWS: Contrastive Learning with hard Attention and Weak Supervision0
Show:102550
← PrevPage 7 of 10Next →

No leaderboard results yet.