SOTAVerified

Visual Commonsense Reasoning

Papers

Showing 2130 of 65 papers

TitleStatusHype
How Vision-Language Tasks Benefit from Large Pre-trained Models: A Survey0
Learning to Correction: Explainable Feedback Generation for Visual Commonsense Reasoning DistractorCode0
Commonsense-T2I Challenge: Can Text-to-Image Generation Models Understand Commonsense?0
ALGO: Object-Grounded Visual Commonsense Reasoning for Open-World Egocentric Action Recognition0
Do Vision-Language Transformers Exhibit Visual Commonsense? An Empirical Study of VCR0
EventLens: Leveraging Event-Aware Pretraining and Cross-modal Linking Enhances Visual Commonsense Reasoning0
ViP-LLaVA: Making Large Multimodal Models Understand Arbitrary Visual PromptsCode0
Improving Vision-and-Language Reasoning via Spatial Relations Modeling0
ViCor: Bridging Visual Understanding and Commonsense Reasoning with Large Language Models0
Discovering Novel Actions from Open World Egocentric Videos with Object-Grounded Visual Commonsense Reasoning0
Show:102550
← PrevPage 3 of 7Next →

No leaderboard results yet.