O1 Replication Journey -- Part 2: Surpassing O1-preview through Simple Distillation, Big Progress or Bitter Lesson? Nov 25, 2024 Hallucination Knowledge Distillation
Code Code Available 7Awesome Multi-modal Object Tracking May 23, 2024 Autonomous Driving Knowledge Distillation
Code Code Available 5A Survey on Knowledge Distillation of Large Language Models Feb 20, 2024 Data Augmentation Knowledge Distillation
Code Code Available 5MobileSAMv2: Faster Segment Anything to Everything Dec 15, 2023 Decoder Knowledge Distillation
Code Code Available 5AlphaPose: Whole-Body Regional Multi-Person Pose Estimation and Tracking in Real-Time Nov 7, 2022 Knowledge Distillation Multi-Person Pose Estimation
Code Code Available 5SAMPart3D: Segment Any Part in 3D Objects Nov 11, 2024 3D Generation 3D Part Segmentation
Code Code Available 4LLM Inference Unveiled: Survey and Roofline Model Insights Feb 26, 2024 Knowledge Distillation Language Modelling
Code Code Available 4Towards Cross-Tokenizer Distillation: the Universal Logit Distillation Loss for LLMs Feb 19, 2024 Knowledge Distillation
Code Code Available 4BitDistiller: Unleashing the Potential of Sub-4-Bit LLMs via Self-Distillation Feb 16, 2024 Knowledge Distillation Quantization
Code Code Available 4Distil-Whisper: Robust Knowledge Distillation via Large-Scale Pseudo Labelling Nov 1, 2023 Hallucination Knowledge Distillation
Code Code Available 4Effective Whole-body Pose Estimation with Two-stages Distillation Jul 29, 2023 2D Human Pose Estimation Knowledge Distillation
Code Code Available 4Vision-Language Models for Vision Tasks: A Survey Apr 3, 2023 Benchmarking Knowledge Distillation
Code Code Available 4Efficient and Generalizable Speaker Diarization via Structured Pruning of Self-Supervised Models Jun 23, 2025 Domain Adaptation GPU
Code Code Available 3Efficient Reasoning Models: A Survey Apr 15, 2025 Knowledge Distillation Model Compression
Code Code Available 3A Survey on Inference Optimization Techniques for Mixture of Experts Models Dec 18, 2024 Computational Efficiency Distributed Computing
Code Code Available 3LLaVA-MoD: Making LLaVA Tiny via MoE Knowledge Distillation Aug 28, 2024 Computational Efficiency Hallucination
Code Code Available 3Compact Language Models via Pruning and Knowledge Distillation Jul 19, 2024 Knowledge Distillation Language Modeling
Code Code Available 3Hydra-MDP: End-to-end Multimodal Planning with Multi-target Hydra-Distillation Jun 11, 2024 Decoder Knowledge Distillation
Code Code Available 3Recurrent Drafter for Fast Speculative Decoding in Large Language Models Mar 14, 2024 Benchmarking Knowledge Distillation
Code Code Available 3PromptKD: Unsupervised Prompt Distillation for Vision-Language Models Mar 5, 2024 Knowledge Distillation Prompt Engineering
Code Code Available 3Logit Standardization in Knowledge Distillation Mar 3, 2024 Knowledge Distillation
Code Code Available 3DistiLLM: Towards Streamlined Distillation for Large Language Models Feb 6, 2024 Instruction Following Knowledge Distillation
Code Code Available 3Generalized Robot 3D Vision-Language Model with Fast Rendering and Pre-Training Vision-Language Alignment Dec 1, 2023 Contrastive Learning Few-Shot Learning
Code Code Available 3ProDiff: Progressive Fast Diffusion Model For High-Quality Text-to-Speech Jul 13, 2022 Denoising GPU
Code Code Available 3CMKD: CNN/Transformer-Based Cross-Model Knowledge Distillation for Audio Classification Mar 13, 2022 Audio Classification Knowledge Distillation
Code Code Available 3N-LTP: An Open-source Neural Language Technology Platform for Chinese Sep 24, 2020 Chinese Word Segmentation Dependency Parsing
Code Code Available 3Semi-Supervised Speech Recognition via Local Prior Matching Feb 24, 2020 Knowledge Distillation Language Modeling
Code Code Available 3Structural Entropy Guided Agent for Detecting and Repairing Knowledge Deficiencies in LLMs May 12, 2025 AI Agent Knowledge Distillation
Code Code Available 2Efficient Multivariate Time Series Forecasting via Calibrated Language Models with Privileged Knowledge Distillation May 4, 2025 Knowledge Distillation Multivariate Time Series Forecasting
Code Code Available 2Distillation-Supervised Convolutional Low-Rank Adaptation for Efficient Image Super-Resolution Apr 15, 2025 Image Super-Resolution Knowledge Distillation
Code Code Available 2Learning Occlusion-Robust Vision Transformers for Real-Time UAV Tracking Apr 12, 2025 Knowledge Distillation
Code Code Available 2SoTA with Less: MCTS-Guided Sample Selection for Data-Efficient Visual Reasoning Self-Improvement Apr 10, 2025 Knowledge Distillation Visual Reasoning
Code Code Available 2Scaling Down Text Encoders of Text-to-Image Diffusion Models Mar 25, 2025 GPU Image Generation
Code Code Available 2A Comprehensive Survey on Knowledge Distillation Mar 15, 2025 Knowledge Distillation Survey
Code Code Available 2LightGen: Efficient Image Generation through Knowledge Distillation and Direct Preference Optimization Mar 11, 2025 GPU Image Generation
Code Code Available 2JL1-CD: A New Benchmark for Remote Sensing Change Detection and a Robust Multi-Teacher Knowledge Distillation Framework Feb 19, 2025 Change Detection Earth Observation
Code Code Available 2Event Stream-based Visual Object Tracking: HDETrack V2 and A High-Definition Benchmark Feb 8, 2025 Knowledge Distillation Object Tracking
Code Code Available 2LightGNN: Simple Graph Neural Network for Recommendation Jan 6, 2025 Computational Efficiency Graph Neural Network
Code Code Available 2Optimizing Edge AI: A Comprehensive Survey on Data, Model, and System Strategies Jan 4, 2025 Edge-computing Knowledge Distillation
Code Code Available 2Learning an Adaptive and View-Invariant Vision Transformer for Real-Time UAV Tracking Dec 28, 2024 Knowledge Distillation Visual Tracking
Code Code Available 2Learnable Prompting SAM-induced Knowledge Distillation for Semi-supervised Medical Image Segmentation Dec 18, 2024 Image Segmentation Knowledge Distillation
Code Code Available 2BiM-VFI: directional Motion Field-Guided Frame Interpolation for Video with Non-uniform Motions Dec 16, 2024 Knowledge Distillation Motion Estimation
Code Code Available 2Wasserstein Distance Rivals Kullback-Leibler Divergence for Knowledge Distillation Dec 11, 2024 image-classification Image Classification
Code Code Available 2BiomedCoOp: Learning to Prompt for Biomedical Vision-Language Models Nov 21, 2024 image-classification Image Classification
Code Code Available 2ScaleKD: Strong Vision Transformers Could Be Excellent Teachers Nov 11, 2024 image-classification Image Classification
Code Code Available 2MiniPLM: Knowledge Distillation for Pre-Training Language Models Oct 22, 2024 Diversity Knowledge Distillation
Code Code Available 2Distillation-Free One-Step Diffusion for Real-World Image Super-Resolution Oct 5, 2024 Image Super-Resolution Knowledge Distillation
Code Code Available 2Learning from Committee: Reasoning Distillation from a Mixture of Teachers with Peer-Review Oct 4, 2024 Knowledge Distillation Logical Reasoning
Code Code Available 2Ruri: Japanese General Text Embeddings Sep 12, 2024 Knowledge Distillation
Code Code Available 2Towards A Generalizable Pathology Foundation Model via Unified Knowledge Distillation Jul 26, 2024 Knowledge Distillation Question Answering
Code Code Available 2