SOTAVerified

Visual Question Answering

MLLM Leaderboard

Papers

Showing 826850 of 2177 papers

TitleStatusHype
Multimodal Preference Data Synthetic Alignment with Reward ModelCode0
COLUMBUS: Evaluating COgnitive Lateral Understanding through Multiple-choice reBUSesCode0
IMAD: IMage-Augmented multi-modal DialogueCode0
Game of Sketches: Deep Recurrent Models of Pictionary-style Word GuessingCode0
Image as a Foreign Language: BEiT Pretraining for All Vision and Vision-Language TasksCode0
Multimodal Explanations: Justifying Decisions and Pointing to the EvidenceCode0
Multi-modal Factorized Bilinear Pooling with Co-Attention Learning for Visual Question AnsweringCode0
Multimodal Compact Bilinear Pooling for Visual Question Answering and Visual GroundingCode0
Multimodal Large Language Models and Tunings: Vision, Language, Sensors, Audio, and BeyondCode0
Fully Authentic Visual Question Answering Dataset from Online CommunitiesCode0
Adapting Visual Question Answering Models for Enhancing Multimodal Community Q&A PlatformsCode0
Language Models Meet Anomaly Detection for Better Interpretability and GeneralizabilityCode0
Right this way: Can VLMs Guide Us to See More to Answer Questions?Code0
Cognitive Visual Commonsense Reasoning Using Dynamic Working MemoryCode0
Multi-Image Visual Question AnsweringCode0
MQA: Answering the Question via Robotic ManipulationCode0
From Images to Textual Prompts: Zero-shot VQA with Frozen Large Language ModelsCode0
Multi-Page Document Visual Question Answering using Self-Attention Scoring MechanismCode0
Modularized Zero-shot VQA with Pre-trained ModelsCode0
FRAMES-VQA: Benchmarking Fine-Tuning Robustness across Multi-Modal Shifts in Visual Question AnsweringCode0
Co-attending Regions and Detections with Multi-modal Multiplicative Embedding for VQACode0
Co-attending Free-form Regions and Detections with Multi-modal Multiplicative Feature Embedding for Visual Question AnsweringCode0
Modulating early visual processing by languageCode0
MM-PoE: Multiple Choice Reasoning via. Process of Elimination using Multi-Modal ModelsCode0
Focal Visual-Text Attention for Visual Question AnsweringCode0
Show:102550
← PrevPage 34 of 88Next →

Benchmark Results

#ModelMetricClaimedVerifiedStatus
1MMCTAgent (GPT-4 + GPT-4V)GPT-4 score74.24Unverified
2Qwen2-VL-72BGPT-4 score74Unverified
3InternVL2.5-78BGPT-4 score72.3Unverified
4GPT-4o +text rationale +IoTGPT-4 score72.2Unverified
5Lyra-ProGPT-4 score71.4Unverified
6GLM-4V-PlusGPT-4 score71.1Unverified
7Phantom-7BGPT-4 score70.8Unverified
8InternVL2.5-38BGPT-4 score68.8Unverified
9InternVL2-26B (SGP, token ratio 64%)GPT-4 score65.6Unverified
10Baichuan-Omni (7B)GPT-4 score65.4Unverified