SOTAVerified

Text Retrieval

Text Retrieval is the task of finding the most text result (such as an answer, paragraph, or passage) given a query (which could be a question, keywords, or any relevant text)

Papers

Showing 51100 of 671 papers

TitleStatusHype
A Replication Study of Dense Passage RetrieverCode2
Efficient Inverted Indexes for Approximate Retrieval over Learned Sparse RepresentationsCode2
Multi-modal Molecule Structure-text Model for Text-based Retrieval and EditingCode2
CLIP-ViP: Adapting Pre-trained Image-Text Model to Video-Language Representation AlignmentCode2
MedCLIP: Contrastive Learning from Unpaired Medical Images and TextCode2
M2-RAAP: A Multi-Modal Recipe for Advancing Adaptation-based Pre-training towards Effective and Efficient Zero-shot Video-text RetrievalCode2
GLAP: General contrastive audio-text pretraining across domains and languagesCode2
AudioSetCaps: An Enriched Audio-Caption Dataset using Automated Generation Pipeline with Large Audio and Language ModelsCode2
Gramian Multimodal Representation Learning and AlignmentCode2
One Trajectory, One Token: Grounded Video Tokenization via Panoptic Sub-object TrajectoryCode2
Audio Retrieval with WavText5K and CLAP TrainingCode1
Audio Retrieval with Natural Language Queries: A Benchmark StudyCode1
Align before Fuse: Vision and Language Representation Learning with Momentum DistillationCode1
A Comparison of Pre-trained Vision-and-Language Models for Multimodal Representation Learning across Medical Images and ReportsCode1
Efficient Token-Guided Image-Text Retrieval with Consistent Multimodal Contrastive TrainingCode1
Efficient Vision-Language Pretraining with Visual Concepts and Hierarchical AlignmentCode1
COM Kitchens: An Unedited Overhead-view Video Dataset as a Vision-Language BenchmarkCode1
A Survey of Medical Vision-and-Language Applications and Their TechniquesCode1
Efficient Medical Vision-Language Alignment Through Adapting Masked Vision ModelsCode1
From Association to Generation: Text-only Captioning by Unsupervised Cross-modal MappingCode1
Frozen in Time: A Joint Video and Image Encoder for End-to-End RetrievalCode1
Dynamic Modality Interaction Modeling for Image-Text RetrievalCode1
FlexiViT: One Model for All Patch SizesCode1
Benchmarking Robustness of Multimodal Image-Text Models under Distribution ShiftCode1
mmRAG: A Modular Benchmark for Retrieval-Augmented Generation over Text, Tables, and Knowledge GraphsCode1
Fine-grained Video-Text Retrieval with Hierarchical Graph ReasoningCode1
Efficiently Teaching an Effective Dense Retriever with Balanced Topic Aware SamplingCode1
DialogCC: An Automated Pipeline for Creating High-Quality Multi-Modal Dialogue DatasetCode1
AdvCLIP: Downstream-agnostic Adversarial Examples in Multimodal Contrastive LearningCode1
DiscoVLA: Discrepancy Reduction in Vision, Language, and Alignment for Parameter-Efficient Video-Text RetrievalCode1
Fine-Tuning LLaMA for Multi-Stage Text RetrievalCode1
Focus, Distinguish, and Prompt: Unleashing CLIP for Efficient and Flexible Scene Text RetrievalCode1
Dense Hierarchical Retrieval for Open-Domain Question AnsweringCode1
Approximate Nearest Neighbor Negative Contrastive Learning for Dense Text RetrievalCode1
FETA: Towards Specializing Foundation Models for Expert Task ApplicationsCode1
A Comprehensive Review of the Video-to-Text ProblemCode1
DecAF: Joint Decoding of Answers and Logical Forms for Question Answering over Knowledge BasesCode1
Text Proxy: Decomposing Retrieval from a 1-to-N Relationship into N 1-to-1 Relationships for Text-Video RetrievalCode1
Bridging Language Gaps in Audio-Text RetrievalCode1
Data-Efficient Multimodal Fusion on a Single GPUCode1
Bridging Video-text Retrieval with Multiple Choice QuestionsCode1
Building an Open-Vocabulary Video CLIP Model with Better Architectures, Optimization and DataCode1
A Prior Instruction Representation Framework for Remote Sensing Image-text RetrievalCode1
A Dense Representation Framework for Lexical and Semantic MatchingCode1
Cross-View Language Modeling: Towards Unified Cross-Lingual Cross-Modal Pre-trainingCode1
CVLUE: A New Benchmark Dataset for Chinese Vision-Language Understanding EvaluationCode1
ArabicaQA: A Comprehensive Dataset for Arabic Question AnsweringCode1
Cross-modal Scene Graph Matching for Relationship-aware Image-Text RetrievalCode1
CLASP: Contrastive Language-Speech Pretraining for Multilingual Multimodal Information RetrievalCode1
A Deep Local and Global Scene-Graph Matching for Image-Text RetrievalCode1
Show:102550
← PrevPage 2 of 14Next →

No leaderboard results yet.