SOTAVerified

Token Reduction

Papers

Showing 2650 of 78 papers

TitleStatusHype
Hierarchical Context Merging: Better Long Context Understanding for Pre-trained LLMsCode1
ALGM: Adaptive Local-then-Global Token Merging for Efficient Semantic Segmentation with Plain Vision TransformersCode1
Token Cropr: Faster ViTs for Quite a Few TasksCode1
Inference Optimal VLMs Need Fewer Visual Tokens and More ParametersCode1
Window Token Concatenation for Efficient Visual Large Language ModelsCode1
ZipR1: Reinforcing Token Sparsity in MLLMs0
AdaFV: Rethinking of Visual-Language alignment for VLM acceleration0
Astraea: A GPU-Oriented Token-wise Acceleration Framework for Video Diffusion Transformers0
AsymRnR: Video Diffusion Transformers Acceleration with Asymmetric Reduction and Restoration0
Cut the Crap: An Economical Communication Pipeline for LLM-based Multi-Agent Systems0
Deploying Foundation Model Powered Agent Services: A Survey0
DRP: Distilled Reasoning Pruning with Skill-aware Step Decomposition for Efficient Large Reasoning Models0
DyMU: Dynamic Merging and Virtual Unmerging for Efficient VLMs0
Dynamic Token Reduction during Generation for Vision Language Models0
EcoSafeRAG: Efficient Security through Context Analysis in Retrieval-Augmented Generation0
Efficient LLaMA-3.2-Vision by Trimming Cross-attended Visual Features0
Efficient Multi-modal Large Language Models via Visual Token Grouping0
FIT-RAG: Black-Box RAG with Factual Information and Token Reduction0
freePruner: A Training-free Approach for Large Multimodal Model Acceleration0
Hypernym Mercury: Token Optimization Through Semantic Field Constriction And Reconstruction From Hypernyms. A New Text Compression Method0
ImagePiece: Content-aware Re-tokenization for Efficient Image Recognition0
Knowing When to Stop: Dynamic Context Cutoff for Large Language Models0
Learning Free Token Reduction for Multi-Modal Large Language Models0
Local Information Matters: Inference Acceleration For Grounded Conversation Generation Models Through Adaptive Local-Aware Token Pruning0
MINT: Mitigating Hallucinations in Large Vision-Language Models via Token Reduction0
Show:102550
← PrevPage 2 of 4Next →

No leaderboard results yet.