SOTAVerified

Medical Visual Question Answering

Papers

Showing 2650 of 97 papers

TitleStatusHype
EHRXQA: A Multi-Modal Question Answering Dataset for Electronic Health Records with Chest X-ray ImagesCode1
Localized Questions in Medical Visual Question AnsweringCode1
Masked Vision and Language Pre-training with Unimodal and Multimodal Contrastive Losses for Medical Visual Question AnsweringCode1
MC-CoT: A Modular Collaborative CoT Framework for Zero-shot Medical-VQA with LLM and MLLM IntegrationCode1
Worse than Random? An Embarrassingly Simple Probing Evaluation of Large Multimodal Models in Medical VQACode1
MedAgentBoard: Benchmarking Multi-Agent Collaboration with Conventional Methods for Diverse Medical TasksCode1
PMC-VQA: Visual Instruction Tuning for Medical Visual Question AnsweringCode1
Gemini Goes to Med School: Exploring the Capabilities of Multimodal Large Language Models on Medical Challenge Problems & HallucinationsCode1
A Survey of Medical Vision-and-Language Applications and Their TechniquesCode1
Rad-ReStruct: A Novel VQA Benchmark and Method for Structured Radiology ReportingCode1
Self-supervised vision-language pretraining for Medical visual question answeringCode1
SLAKE: A Semantically-Labeled Knowledge-Enhanced Dataset for Medical Visual Question AnsweringCode1
STLLaVA-Med: Self-Training Large Language and Vision Assistant for Medical Question-AnsweringCode1
ClinKD: Cross-Modal Clinical Knowledge Distiller For Multi-Task Medical ImagesCode0
ViLMedic: a framework for research at the intersection of vision and language in medical AICode0
A Dual-Attention Learning Network with Word and Sentence Embedding for Medical Visual Question AnsweringCode0
FEDMEKI: A Benchmark for Scaling Medical Foundation Models via Federated Knowledge InjectionCode0
Kvasir-VQA: A Text-Image Pair GI Tract DatasetCode0
Hallucination Benchmark in Medical Visual Question AnsweringCode0
Hierarchical Deep Multi-modal Network for Medical Visual Question AnsweringCode0
Kvasir-VQA-x1: A Multimodal Dataset for Medical Reasoning and Robust MedVQA in Gastrointestinal EndoscopyCode0
Med-PMC: Medical Personalized Multi-modal Consultation with a Proactive Ask-First-Observe-Next ParadigmCode0
Targeted Visual Prompting for Medical Visual Question AnsweringCode0
Does CLIP Benefit Visual Question Answering in the Medical Domain as Much as it Does in the General Domain?0
ZALM3: Zero-Shot Enhancement of Vision-Language Alignment via In-Context Information in Multi-Turn Multimodal Medical Dialogue0
Show:102550
← PrevPage 2 of 4Next →

No leaderboard results yet.