SOTAVerified

Referring expression generation

Generate referring expressions

Papers

Showing 150 of 84 papers

TitleStatusHype
Mini-Gemini: Mining the Potential of Multi-modality Vision Language ModelsCode7
MiniGPT-v2: large language model as a unified interface for vision-language multi-task learningCode7
Improved Baselines with Visual Instruction TuningCode6
Visual Instruction TuningCode6
Efficient Multimodal Learning from Data-centric PerspectiveCode5
LLaVA-Med: Training a Large Language-and-Vision Assistant for Biomedicine in One DayCode4
MobileVLM : A Fast, Strong and Open Vision Language Assistant for Mobile DevicesCode3
Frontiers in Intelligent ColonoscopyCode2
Elysium: Exploring Object-level Perception in Videos via MLLMCode2
GLaMM: Pixel Grounding Large Multimodal ModelCode2
Uni-Med: A Unified Medical Generalist Foundation Model For Multi-Task Learning Via Connector-MoECode1
Multi-modal Instruction Tuned LLMs with Fine-grained Visual PerceptionCode1
Kosmos-2: Grounding Multimodal Large Language Models to the WorldCode1
Modeling Context in Referring ExpressionsCode1
Vision-Language Models Are Not Pragmatically Competent in Referring Expression GenerationCode0
Grounding Language in Multi-Perspective Referential CommunicationCode0
Referring Expression Generation in Visually Grounded Dialogue with Discourse-aware Comprehension GuidingCode0
Resilience through Scene Context in Visual Referring Expression GenerationCode0
Intrinsic Task-based Evaluation for Referring Expression Generation0
Enhancing Visual Grounding and Generalization: A Multi-Task Cycle Training Approach for Vision-Language ModelsCode0
Collecting Visually-Grounded Dialogue with A Game Of SortsCode0
Whether you can locate or not? Interactive Referring Expression GenerationCode0
DisCLIP: Open-Vocabulary Referring Expression Generation0
Pento-DIARef: A Diagnostic Dataset for Learning the Incremental Algorithm for Referring Expression Generation from ExamplesCode0
Assessing Neural Referential Form Selectors on a Realistic Multilingual Dataset0
Referring Expressions with Rational Speech Act Framework: A Probabilistic Approach0
Non-neural Models Matter: A Re-evaluation of Neural Referring Expression Generation Systems0
Using Referring Expression Generation to Model Literary Style0
Decoupling Pragmatics: Discriminative Decoding for Referring Expression Generation0
What can Neural Referential Form Selectors Learn?0
Enriching the E2E datasetCode0
Perspective-corrected Spatial Referring Expression Generation for Human-Robot Interaction0
Visual Question Answering based on Local-Scene-Aware Referring Expression Generation0
Improving the Naturalness and Diversity of Referring Expression Generation models using Minimum Risk Training0
OMEGA : A probabilistic approach to referring expression generation in a virtual environment0
Referring to what you know and do not know: Making Referring Expression Generation Models Generalize To Unseen Entities0
Generating Quantified Referring Expressions through Attention-Driven Incremental Perception0
CoNAN: A Complementary Neighboring-based Attention Network for Referring Expression Generation0
Lessons from Computational Modelling of Reference Production in Mandarin and English0
Fuzzy Logic for Vagueness Management in Referring Expression Generation0
Toward Forgetting-Sensitive Referring Expression Generationfor Integrated Robot Architectures0
Informativity in Image Captions vs. Referring Expressions0
MuDoCo: Corpus for Multidomain Coreference Resolution and Referring Expression Generation0
A case study on context-bound referring expression generation0
Improving Quality and Efficiency in Plan-based Neural Data-to-Text GenerationCode0
Referring Expression Generation Using Entity ProfilesCode0
Augmenting Robot Knowledge Consultants with Distributed Short Term Memory0
Adapting Descriptions of People to the Point of View of a Moving Observer0
Enriching the WebNLG corpusCode0
Decoding Strategies for Neural Referring Expression Generation0
Show:102550
← PrevPage 1 of 2Next →

Benchmark Results

#ModelMetricClaimedVerifiedStatus
1ColonGPT (w/ LoRA, w/o extra data)Accuray99.96Unverified
2LLaVA-v1.5 (w/ LoRA, w/ extra data)Accuray99.32Unverified
3LLaVA-Med-v1.5 (w/ LoRA, w/o extra data)Accuray99.3Unverified
4MGM-2B (w/o LoRA, w/ extra data)Accuray98.75Unverified
5LLaVA-v1.5 (w/ LoRA, w/o extra data)Accuray98.58Unverified
6MGM-2B (w/o LoRA, w/o extra data)Accuray98.17Unverified
7MobileVLM-1.7B (w/ LoRA, w/ extra data)Accuray97.87Unverified
8MobileVLM-1.7B (w/o LoRA, w/ extra data)Accuray97.78Unverified
9LLaVA-Med-v1.0 (w/o LoRA, w/o extra data)Accuray97.74Unverified
10LLaVA-Med-v1.0 (w/o LoRA, w/ extra data)Accuray97.35Unverified
#ModelMetricClaimedVerifiedStatus
1LLaVA-Med-v1.5 (w/ LoRA, w/ extra data)Accuray70Unverified
2LLaVA-v1 (w/ LoRA, w/ extra data)Accuray46.85Unverified