SOTAVerified

document understanding

Document understanding involves document classification, layout analysis, information extraction, and DocQA.

Papers

Showing 150 of 309 papers

TitleStatusHype
Qwen2.5-VL Technical ReportCode11
DocLayout-YOLO: Enhancing Document Layout Analysis through Diverse Synthetic Data and Global-to-Local Adaptive PerceptionCode9
GLM-4.1V-Thinking: Towards Versatile Multimodal Reasoning with Scalable Reinforcement LearningCode7
ColPali: Efficient Document Retrieval with Vision Language ModelsCode7
Mini-Monkey: Alleviating the Semantic Sawtooth Effect for Lightweight MLLMs via Complementary Image PyramidCode5
Focus Anywhere for Fine-grained Multi-page Document UnderstandingCode5
TextMonkey: An OCR-Free Large Multimodal Model for Understanding DocumentCode5
LLMMapReduce: Simplified Long-Sequence Processing using Large Language ModelsCode4
MDocAgent: A Multi-Modal Multi-Agent Framework for Document UnderstandingCode3
Oryx MLLM: On-Demand Spatial-Temporal Understanding at Arbitrary ResolutionCode3
INTERS: Unlocking the Power of Large Language Models in Search with Instruction TuningCode3
Unifying Vision, Text, and Layout for Universal Document ProcessingCode3
OCR-free Document Understanding TransformerCode3
AIN: The Arabic INclusive Large Multimodal ModelCode2
Arabic-Nougat: Fine-Tuning Vision Transformers for Arabic OCR and Markdown ExtractionCode2
PDF-WuKong: A Large Multimodal Model for Efficient Long PDF Reading with End-to-End Sparse SamplingCode2
One missing piece in Vision and Language: A Survey on Comics UnderstandingCode2
A Bounding Box is Worth One Token: Interleaving Layout and Text in a Large Language Model for Document UnderstandingCode2
MMLongBench-Doc: Benchmarking Long-context Document Understanding with VisualizationsCode2
Visually Guided Generative Text-Layout Pre-training for Document IntelligenceCode2
InstructDoc: A Dataset for Zero-Shot Generalization of Visual Document Understanding with InstructionsCode2
Delivering Document Conversion as a Cloud Service with High Throughput and ResponsivenessCode2
LiLT: A Simple yet Effective Language-Independent Layout Transformer for Structured Document UnderstandingCode2
ICDAR 2021 Competition on Scientific Literature ParsingCode2
SimpleDoc: Multi-Modal Document Understanding with Dual-Cue Page Retrieval and Iterative RefinementCode1
LEMONADE: A Large Multilingual Expert-Annotated Abstractive Event Dataset for the Real WorldCode1
ARB: A Comprehensive Arabic Multimodal Reasoning BenchmarkCode1
Adaptive Markup Language Generation for Contextually-Grounded Visual Document UnderstandingCode1
FRAG: Frame Selection Augmented Generation for Long Video and Long Document UnderstandingCode1
Ocean-OCR: Towards General OCR Application via a Vision-Language ModelCode1
DocLayLLM: An Efficient Multi-modal Extension of Large Language Models for Text-rich Document UnderstandingCode1
Docopilot: Improving Multimodal Models for Document-Level UnderstandingCode1
LongDocURL: a Comprehensive Multimodal Long Document Benchmark Integrating Understanding, Reasoning, and LocatingCode1
Typhoon 2: A Family of Open Text and Multimodal Thai Large Language ModelsCode1
CAMEL-Bench: A Comprehensive Arabic LMM BenchmarkCode1
Modeling Layout Reading Order as Ordering Relations for Visually-rich Document UnderstandingCode1
DocLayLLM: An Efficient and Effective Multi-modal Extension of Large Language Models for Text-rich Document UnderstandingCode1
VisFocus: Prompt-Guided Vision Encoders for OCR-Free Dense Document UnderstandingCode1
DANIEL: A fast Document Attention Network for Information Extraction and Labelling of handwritten documentsCode1
Leveraging Visual Tokens for Extended Text Contexts in Multi-Modal LearningCode1
Hierarchical Multimodal Pre-training for Visually Rich Webpage UnderstandingCode1
On the Affinity, Rationality, and Diversity of Hierarchical Topic ModelingCode1
WordScape: a Pipeline to extract multilingual, visually rich Documents with Layout Annotations from Web Crawl DataCode1
Privacy-Aware Document Visual Question AnsweringCode1
Towards Improving Document Understanding: An Exploration on Text-Grounding via MLLMsCode1
DocTrack: A Visually-Rich Document Dataset Really Aligned with Human Eye Movement for Machine ReadingCode1
Enhancing Visually-Rich Document Understanding via Layout Structure ModelingCode1
DocumentCLIP: Linking Figures and Main Body Text in Reflowed DocumentsCode1
DocFormerv2: Local Features for Document UnderstandingCode1
PaLI-X: On Scaling up a Multilingual Vision and Language ModelCode1
Show:102550
← PrevPage 1 of 7Next →

No leaderboard results yet.