SOTAVerified

Token Reduction

Papers

Showing 5175 of 78 papers

TitleStatusHype
Token Cropr: Faster ViTs for Quite a Few TasksCode1
Accelerating Multimodal Large Language Models by Searching Optimal Vision Token Reduction0
Efficient Multi-modal Large Language Models via Visual Token Grouping0
Rethinking Token Reduction in MLLMs: Towards a Unified Paradigm for Training-Free Acceleration0
freePruner: A Training-free Approach for Large Multimodal Model Acceleration0
Inference Optimal VLMs Need Fewer Visual Tokens and More ParametersCode1
LongVU: Spatiotemporal Adaptive Compression for Long Video-Language UnderstandingCode3
Rethinking Token Reduction for State Space ModelsCode1
PAR: Prompt-Aware Token Reduction Method for Efficient Large Multimodal Models0
FastAdaSP: Multitask-Adapted Efficient Inference for Large Speech Language ModelCode1
Cut the Crap: An Economical Communication Pipeline for LLM-based Multi-Agent Systems0
Discovering the Gems in Early Layers: Accelerating Long-Context LLMs with 1000x Input Token ReductionCode2
Less is More: A Simple yet Effective Token Reduction Method for Efficient Multi-modal LLMsCode1
Vote&Mix: Plug-and-Play Token Reduction for Efficient Vision Transformer0
Bridging Local Details and Global Context in Text-Attributed GraphsCode1
ALGM: Adaptive Local-then-Global Token Merging for Efficient Semantic Segmentation with Plain Vision TransformersCode1
Hierarchical Context Merging: Better Long Context Understanding for Pre-trained LLMsCode1
LLaVA-PruMerge: Adaptive Token Reduction for Efficient Large Multimodal ModelsCode2
FIT-RAG: Black-Box RAG with Factual Information and Token Reduction0
HaltingVT: Adaptive Token Halting Transformer for Efficient Video RecognitionCode0
TPC-ViT: Token Propagation Controller for Efficient Vision Transformer0
Which Tokens to Use? Investigating Token Reduction in Vision TransformersCode1
Content-aware Token Sharing for Efficient Semantic Segmentation with Vision TransformersCode1
PuMer: Pruning and Merging Tokens for Efficient Vision Language ModelsCode1
Selective Structured State-Spaces for Long-Form Video Understanding0
Show:102550
← PrevPage 3 of 4Next →

No leaderboard results yet.