SOTAVerified

Hard Attention

Papers

Showing 51100 of 100 papers

TitleStatusHype
Specialized Transformers: Faster, Smaller and more Accurate NLP Models0
Text as Environment: A Deep Reinforcement Learning Text Readability Assessment Model0
Theoretical Limitations of Self-Attention in Neural Sequence Models0
Transformers as Transducers0
Transformers in Uniform TC^00
Unique Hard Attention: A Tale of Two Sides0
Upper, Middle and Lower Region Learning for Facial Action Unit Detection0
Video Violence Recognition and Localization Using a Semi-Supervised Hard Attention Model0
Word Representation Models for Morphologically Rich Languages in Neural Machine Translation0
You Only Need One Model for Open-domain Question Answering0
NoPE: The Counting Power of Transformers with No Positional Encodings0
Achieving Explainability in a Visual Hard Attention Model through Content Prediction0
A Differentiable Self-disambiguated Sense Embedding Model via Scaled Gumbel Softmax0
AMR Parsing with Action-Pointer Transformer0
An Exploration of Neural Sequence-to-Sequence Architectures for Automatic Post-Editing0
A study of latent monotonic attention variants0
AttentionDrop: A Novel Regularization Method for Transformer Models0
Average-Hard Attention Transformers are Constant-Depth Uniform Threshold Circuits0
Characterizing the Expressivity of Transformer Language Models0
CLAWS: Contrastive Learning with hard Attention and Weak Supervision0
Comparison of different Unique hard attention transformer models by the formal languages they can recognize0
Continual Diffusion with STAMINA: STack-And-Mask INcremental Adapters0
DanHAR: Dual Attention Network For Multimodal Human Activity Recognition Using Wearable Sensors0
Deep Pneumonia: Attention-Based Contrastive Learning for Class-Imbalanced Pneumonia Lesion Recognition in Chest X-rays0
Effect of choice of probability distribution, randomness, and search methods for alignment modeling in sequence-to-sequence text-to-speech synthesis using hard alignment0
Ehrenfeucht-Haussler Rank and Chain of Thought0
Emergence of Fixational and Saccadic Movements in a Multi-Level Recurrent Attention Model for Vision0
Exact Expressive Power of Transformers with Padding0
Language-Guided Reinforcement Learning for Hard Attention in Few-Shot Learning0
Extractive Adversarial Networks: High-Recall Explanations for Identifying Personal Attacks in Social Media Posts0
Formal Language Recognition by Hard Attention Transformers: Perspectives from Circuit Complexity0
Generative Adversarial Networks Based on Collaborative Learning and Attention Mechanism for Hyperspectral Image Classification0
GQHAN: A Grover-inspired Quantum Hard Attention Network0
Graph Decoupling Attention Markov Networks for Semi-supervised Graph Node Classification0
Hard Attention Control By Mutual Information Maximization0
Hierarchical Memory Networks0
Hierarchical Multi-scale Attention Networks for Action Recognition0
Improved Attention Models for Memory Augmented Neural Network Adaptive Controllers0
Learning deep graph matching with channel-independent embedding and Hungarian attention0
Learning Hard Alignments with Variational Inference0
Logical Languages Accepted by Transformer Encoders with Hard Attention0
Look Harder: A Neural Machine Translation Model with Hard Attention0
Lower Bounds for Chain-of-Thought Reasoning in Hard-Attention Transformers0
Masked Hard-Attention Transformers Recognize Exactly the Star-Free Languages0
MESAHA-Net: Multi-Encoders based Self-Adaptive Hard Attention Network with Maximum Intensity Projections for Lung Nodule Segmentation in CT Scan0
Dual Attention Model with Reinforcement Learning for Classification of Histology Whole-Slide Images0
Multimodal Emergent Fake News Detection via Meta Neural Process Networks0
MultiResolution Attention Extractor for Small Object Detection0
Multi-View Unsupervised Image Generation with Cross Attention Guidance0
Near-Optimal Glimpse Sequences for Improved Hard Attention Neural Network Training0
Show:102550
← PrevPage 2 of 2Next →

No leaderboard results yet.