SOTAVerified

Referring Expression Comprehension

Papers

Showing 5160 of 167 papers

TitleStatusHype
Coarse-to-Fine Vision-Language Pre-training with Fusion in the BackboneCode1
PEVL: Position-enhanced Pre-training and Prompt Tuning for Vision-language ModelsCode1
A Survivor in the Era of Large-Scale Pretraining: An Empirical Study of One-Stage Referring Expression ComprehensionCode1
ReCLIP: A Strong Zero-Shot Baseline for Referring Expression ComprehensionCode1
SeqTR: A Simple yet Universal Network for Visual GroundingCode1
Bottom Up Top Down Detection Transformers for Language Grounding in Images and Point CloudsCode1
Referring Transformer: A One-step Approach to Multi-task Visual GroundingCode1
MDETR -- Modulated Detection for End-to-End Multi-Modal UnderstandingCode1
TransVG: End-to-End Visual Grounding with TransformersCode1
Unifying Vision-and-Language Tasks via Text GenerationCode1
Show:102550
← PrevPage 6 of 17Next →

No leaderboard results yet.