SOTAVerified

MME

MME is a comprehensive evaluation benchmark for multimodal large language models. It measures both perception and cognition abilities on a total of 14 subtasks, including existence, count, position, color, poster, celebrity, scene, landmark, artwork, OCR, commonsense reasoning, numerical calculation, text translation, and code reasoning.

Papers

Showing 9195 of 95 papers

TitleStatusHype
Scalable K-Medoids via True Error Bound and Familywise Bandits0
Silkie: Preference Distillation for Large Visual Language Models0
Temporal Preference Optimization for Long-Form Video Understanding0
Temporal Reasoning Transfer from Text to Video0
The Use of Symmetry for Models with Variable-size Variables0
Show:102550
← PrevPage 10 of 10Next →

No leaderboard results yet.