O1 Replication Journey -- Part 2: Surpassing O1-preview through Simple Distillation, Big Progress or Bitter Lesson? Nov 25, 2024 Hallucination Knowledge Distillation
Code Code Available 75 AlphaPose: Whole-Body Regional Multi-Person Pose Estimation and Tracking in Real-Time Nov 7, 2022 Knowledge Distillation Multi-Person Pose Estimation
Code Code Available 55 A Survey on Knowledge Distillation of Large Language Models Feb 20, 2024 Data Augmentation Knowledge Distillation
Code Code Available 55 MobileSAMv2: Faster Segment Anything to Everything Dec 15, 2023 Decoder Knowledge Distillation
Code Code Available 55 Awesome Multi-modal Object Tracking May 23, 2024 Autonomous Driving Knowledge Distillation
Code Code Available 55 Towards Cross-Tokenizer Distillation: the Universal Logit Distillation Loss for LLMs Feb 19, 2024 Knowledge Distillation
Code Code Available 45 Vision-Language Models for Vision Tasks: A Survey Apr 3, 2023 Benchmarking Knowledge Distillation
Code Code Available 45 SAMPart3D: Segment Any Part in 3D Objects Nov 11, 2024 3D Generation 3D Part Segmentation
Code Code Available 45 Effective Whole-body Pose Estimation with Two-stages Distillation Jul 29, 2023 2D Human Pose Estimation Knowledge Distillation
Code Code Available 45 Distil-Whisper: Robust Knowledge Distillation via Large-Scale Pseudo Labelling Nov 1, 2023 Hallucination Knowledge Distillation
Code Code Available 45 LLM Inference Unveiled: Survey and Roofline Model Insights Feb 26, 2024 Knowledge Distillation Language Modelling
Code Code Available 45 BitDistiller: Unleashing the Potential of Sub-4-Bit LLMs via Self-Distillation Feb 16, 2024 Knowledge Distillation Quantization
Code Code Available 45 Generalized Robot 3D Vision-Language Model with Fast Rendering and Pre-Training Vision-Language Alignment Dec 1, 2023 Contrastive Learning Few-Shot Learning
Code Code Available 35 Logit Standardization in Knowledge Distillation Mar 3, 2024 Knowledge Distillation
Code Code Available 35 Efficient Reasoning Models: A Survey Apr 15, 2025 Knowledge Distillation Model Compression
Code Code Available 35 Hydra-MDP: End-to-end Multimodal Planning with Multi-target Hydra-Distillation Jun 11, 2024 Decoder Knowledge Distillation
Code Code Available 35 PromptKD: Unsupervised Prompt Distillation for Vision-Language Models Mar 5, 2024 Knowledge Distillation Prompt Engineering
Code Code Available 35 DistiLLM: Towards Streamlined Distillation for Large Language Models Feb 6, 2024 Instruction Following Knowledge Distillation
Code Code Available 35 Efficient and Generalizable Speaker Diarization via Structured Pruning of Self-Supervised Models Jun 23, 2025 Domain Adaptation GPU
Code Code Available 35 Compact Language Models via Pruning and Knowledge Distillation Jul 19, 2024 Knowledge Distillation Language Modeling
Code Code Available 35 CMKD: CNN/Transformer-Based Cross-Model Knowledge Distillation for Audio Classification Mar 13, 2022 Audio Classification Knowledge Distillation
Code Code Available 35 Recurrent Drafter for Fast Speculative Decoding in Large Language Models Mar 14, 2024 Benchmarking Knowledge Distillation
Code Code Available 35 ProDiff: Progressive Fast Diffusion Model For High-Quality Text-to-Speech Jul 13, 2022 Denoising GPU
Code Code Available 35 N-LTP: An Open-source Neural Language Technology Platform for Chinese Sep 24, 2020 Chinese Word Segmentation Dependency Parsing
Code Code Available 35 LLaVA-MoD: Making LLaVA Tiny via MoE Knowledge Distillation Aug 28, 2024 Computational Efficiency Hallucination
Code Code Available 35 A Survey on Inference Optimization Techniques for Mixture of Experts Models Dec 18, 2024 Computational Efficiency Distributed Computing
Code Code Available 35 Semi-Supervised Speech Recognition via Local Prior Matching Feb 24, 2020 Knowledge Distillation Language Modeling
Code Code Available 35 Focal Loss for Dense Object Detection Aug 7, 2017 2D Object Detection Dense Object Detection
Code Code Available 25 Anomaly Detection via Reverse Distillation from One-Class Embedding Jan 26, 2022 Anomaly Classification
Code Code Available 25 From Instance Training to Instruction Learning: Task Adapters Generation from Instructions Jun 18, 2024 Knowledge Distillation
Code Code Available 25 Event Stream-based Visual Object Tracking: A High-Resolution Benchmark Dataset and A Novel Baseline Sep 26, 2023 Knowledge Distillation Object Tracking
Code Code Available 25 Event Stream-based Visual Object Tracking: HDETrack V2 and A High-Definition Benchmark Feb 8, 2025 Knowledge Distillation Object Tracking
Code Code Available 25 Faster Diffusion: Rethinking the Role of the Encoder for Diffusion Model Inference Dec 15, 2023 Decoder Denoising
Code Code Available 25 Learning from Committee: Reasoning Distillation from a Mixture of Teachers with Peer-Review Oct 4, 2024 Knowledge Distillation Logical Reasoning
Code Code Available 25 EPTQ: Enhanced Post-Training Quantization via Hessian-guided Network-wise Optimization Sep 20, 2023 Knowledge Distillation object-detection
Code Code Available 25 Efficient Large-scale Audio Tagging via Transformer-to-CNN Knowledge Distillation Nov 9, 2022 Audio Classification Audio Tagging
Code Code Available 25 ECLIPSE: Efficient Continual Learning in Panoptic Segmentation with Visual Prompt Tuning Mar 29, 2024 Continual Learning Continual Panoptic Segmentation
Code Code Available 25 Efficient Multivariate Time Series Forecasting via Calibrated Language Models with Privileged Knowledge Distillation May 4, 2025 Knowledge Distillation Multivariate Time Series Forecasting
Code Code Available 25 ESP-MedSAM: Efficient Self-Prompting SAM for Universal Domain-Generalized Medical Image Segmentation Jul 19, 2024 Decoder Image Segmentation
Code Code Available 25 Distillation-Supervised Convolutional Low-Rank Adaptation for Efficient Image Super-Resolution Apr 15, 2025 Image Super-Resolution Knowledge Distillation
Code Code Available 25 Diffusion Time-step Curriculum for One Image to 3D Generation Apr 6, 2024 3D Generation Image to 3D
Code Code Available 25 Decoupled Knowledge Distillation Mar 16, 2022 image-classification Image Classification
Code Code Available 25 Distillation-Free One-Step Diffusion for Real-World Image Super-Resolution Oct 5, 2024 Image Super-Resolution Knowledge Distillation
Code Code Available 25 DOT: A Distillation-Oriented Trainer Jul 17, 2023 Knowledge Distillation
Code Code Available 25 ConDistFL: Conditional Distillation for Federated Learning from Partially Annotated Data Aug 8, 2023 Federated Learning Knowledge Distillation
Code Code Available 25 OBSeg: Accurate and Fast Instance Segmentation Framework Using Segmentation Foundation Models with Oriented Bounding Box Prompts Jan 16, 2024 Amodal Instance Segmentation Instance Segmentation
Code Code Available 25 Cross-Image Relational Knowledge Distillation for Semantic Segmentation Apr 14, 2022 Knowledge Distillation Segmentation
Code Code Available 25 CoLaDa: A Collaborative Label Denoising Framework for Cross-lingual Named Entity Recognition May 24, 2023 Denoising Knowledge Distillation
Code Code Available 25 Data-Free Knowledge Distillation for Deep Neural Networks Oct 19, 2017 Data-free Knowledge Distillation Knowledge Distillation
Code Code Available 25 Dual-Space Knowledge Distillation for Large Language Models Jun 25, 2024 Instruction Following Knowledge Distillation
Code Code Available 25