Visual Question Answering
MLLM Leaderboard
Papers
Showing 1–10 of 2177 papers
All datasetsMM-VetViP-BenchVQA v2 test-devBenchLMMMMBenchV*benchVQA v2 valMSRVTT-QAVQA v2 test-stdMMHal-BenchMSVD-QAPlotQA-D1
Benchmark Results
| # | Model | Metric | Claimed | Verified | Status |
|---|---|---|---|---|---|
| 1 | GPT-4V-turbo-detail:high (Visual Prompt) | GPT-4 score (bbox) | 60.7 | — | Unverified |
| 2 | GPT-4V-turbo-detail:low (Visual Prompt) | GPT-4 score (bbox) | 52.8 | — | Unverified |
| 3 | LLaVA-NeXT-Inst-IT-Qwen2-7B (Visual Prompt | GPT-4 score (bbox) | 50.5 | — | Unverified |
| 4 | ViP-LLaVA-13B (Visual Prompt) | GPT-4 score (bbox) | 48.3 | — | Unverified |
| 5 | LLaVA-1.5-13B (Coordinates) | GPT-4 score (bbox) | 47.1 | — | Unverified |
| 6 | Qwen-VL-Chat (Coordinates) | GPT-4 score (bbox) | 45.3 | — | Unverified |
| 7 | LLaVA-NeXT-Inst-IT-Vicuna-7B (Visual Prompt | GPT-4 score (bbox) | 45.1 | — | Unverified |
| 8 | LLaVA-1.5-13B (Visual Prompt) | GPT-4 score (bbox) | 41.8 | — | Unverified |
| 9 | Qwen-VL-Chat (Visual Prompt) | GPT-4 score (bbox) | 39.2 | — | Unverified |
| 10 | InstructBLIP-13B (Visual Prompt) | GPT-4 score (bbox) | 35.8 | — | Unverified |