A Survey on Model Compression for Large Language Models Aug 15, 2023 Benchmarking Knowledge Distillation
— Unverified 00 A Survey on Recent Teacher-student Learning Studies Apr 10, 2023 Knowledge Distillation Survey
— Unverified 00 A Survey on Symbolic Knowledge Distillation of Large Language Models Jul 12, 2024 Knowledge Distillation Survey
— Unverified 00 A Survey on Transformer Compression Feb 5, 2024 Knowledge Distillation Mamba
— Unverified 00 Asymmetric Decision-Making in Online Knowledge Distillation:Unifying Consensus and Divergence Mar 9, 2025 Decision Making Knowledge Distillation
— Unverified 00 ADPS: Asymmetric Distillation Post-Segmentation for Image Anomaly Detection Oct 19, 2022 Anomaly Detection Anomaly Localization
— Unverified 00 Asymmetric Image Retrieval with Cross Model Compatible Ensembles Mar 30, 2023 Diversity Face Recognition
— Unverified 00 Asymmetric Temperature Scaling Makes Larger Networks Teach Well Again Oct 10, 2022 Knowledge Distillation
— Unverified 00 Asynchronous Convergence in Multi-Task Learning via Knowledge Distillation from Converged Tasks Jul 1, 2022 Knowledge Distillation Multi-Task Learning
— Unverified 00 Edge Bias in Federated Learning and its Solution by Buffered Knowledge Distillation Oct 20, 2020 Federated Learning Knowledge Distillation
— Unverified 00 A Technical Study into Small Reasoning Language Models Jun 16, 2025 Code Generation Computational Efficiency
— Unverified 00 A Theoretical Analysis of Soft-Label vs Hard-Label Training in Neural Networks Dec 12, 2024 Binary Classification Knowledge Distillation
— Unverified 00 A Transformer-in-Transformer Network Utilizing Knowledge Distillation for Image Recognition Feb 24, 2025 image-classification Image Classification
— Unverified 00 Attention-Guided Answer Distillation for Machine Reading Comprehension Aug 23, 2018 Knowledge Distillation Machine Reading Comprehension
— Unverified 00 Attention-guided Feature Distillation for Semantic Segmentation Mar 8, 2024 Knowledge Distillation Segmentation
— Unverified 00 Attention is all you need for boosting graph convolutional neural network Mar 10, 2024 All Knowledge Distillation
— Unverified 00 AttentionLite: Towards Efficient Self-Attention Models for Vision Dec 21, 2020 Knowledge Distillation
— Unverified 00 MKD: a Multi-Task Knowledge Distillation Approach for Pretrained Language Models Nov 9, 2019 Knowledge Distillation Multi-Task Learning
— Unverified 00 Audio-Oriented Multimodal Machine Comprehension: Task, Dataset and Model Jul 4, 2021 Knowledge Distillation Machine Reading Comprehension
— Unverified 00 Audio Representation Learning by Distilling Video as Privileged Information Feb 6, 2023 Emotion Recognition Knowledge Distillation
— Unverified 00 Augmentation with Projection: Towards an Effective and Efficient Data Augmentation Paradigm for Distillation Oct 21, 2022 Data Augmentation Diversity
— Unverified 00 Augmenting Knowledge Distillation With Peer-To-Peer Mutual Learning For Model Compression Oct 21, 2021 Knowledge Distillation Model Compression
— Unverified 00 A Unified Compression Framework for Efficient Speech-Driven Talking-Face Generation Apr 2, 2023 Face Generation Knowledge Distillation
— Unverified 00 A Unified Framework for Continual Learning and Unlearning Aug 21, 2024 Continual Learning Knowledge Distillation
— Unverified 00 A Unified Knowledge-Distillation and Semi-Supervised Learning Framework to Improve Industrial Ads Delivery Systems Feb 5, 2025 Knowledge Distillation
— Unverified 00 A Unified Knowledge Distillation Framework for Deep Directed Graphical Models Sep 29, 2021 Continual Learning Federated Learning
— Unverified 00 AutoADR: Automatic Model Design for Ad Relevance Oct 14, 2020 AutoML Knowledge Distillation
— Unverified 00 AutoDistil: Few-shot Task-agnostic Neural Architecture Search for Distilling Large Language Models Jan 29, 2022 Inductive Bias Knowledge Distillation
— Unverified 00 AutoDistill: an End-to-End Framework to Explore and Distill Hardware-Efficient Language Models Jan 21, 2022 Bayesian Optimization Knowledge Distillation
— Unverified 00 AUTOKD: Automatic Knowledge Distillation Into A Student Architecture Family Nov 5, 2021 Bayesian Optimization Knowledge Distillation
— Unverified 00 Automated Channel Pruning with Learned Importance Sep 29, 2021 Denoising GPU
— Unverified 00 Automated Graph Self-supervised Learning via Multi-teacher Knowledge Distillation Oct 5, 2022 Graph Representation Learning Knowledge Distillation
— Unverified 00 Automatic Block-wise Pruning with Auxiliary Gating Structures for Deep Convolutional Neural Networks May 7, 2022 Knowledge Distillation Model Compression
— Unverified 00 Automatic Mixed-Precision Quantization Search of BERT Dec 30, 2021 Knowledge Distillation Model Compression
— Unverified 00 AUTOSUMM: Automatic Model Creation for Text Summarization Nov 1, 2021 Abstractive Text Summarization Deep Learning
— Unverified 00 A vision transformer-based framework for knowledge transfer from multi-modal to mono-modal lymphoma subtyping models Aug 2, 2023 Knowledge Distillation Transfer Learning
— Unverified 00 Aware of the History: Trajectory Forecasting with the Local Behavior Data Jul 20, 2022 Knowledge Distillation Prediction
— Unverified 00 AWF: Adaptive Weight Fusion for Enhanced Class Incremental Semantic Segmentation Sep 13, 2024 Class-Incremental Semantic Segmentation Knowledge Distillation
— Unverified 00 BabyHGRN: Exploring RNNs for Sample-Efficient Training of Language Models Dec 20, 2024 Knowledge Distillation Language Modeling
— Unverified 00 Background Adaptation with Residual Modeling for Exemplar-Free Class-Incremental Semantic Segmentation Jul 13, 2024 Class-Incremental Semantic Segmentation Exemplar-Free
— Unverified 00 Knowledge Distillation for Human Action Anticipation Apr 9, 2019 Action Anticipation Action Recognition
— Unverified 00 Baidu Neural Machine Translation Systems for WMT19 Aug 1, 2019 Data Augmentation Domain Adaptation
— Unverified 00 Balance Divergence for Knowledge Distillation Jan 14, 2025 image-classification Image Classification
— Unverified 00 Balanced softmax cross-entropy for incremental learning with and without memory Mar 23, 2021 class-incremental learning Class Incremental Learning
— Unverified 00 Balancing Cost and Benefit with Tied-Multi Transformers Feb 20, 2020 Decoder Knowledge Distillation
— Unverified 00 A predictive machine learning force field framework for liquid electrolyte development Apr 10, 2024 Knowledge Distillation
— Unverified 00 BanglaEmbed: Efficient Sentence Embedding Models for a Low-Resource Language Using Cross-Lingual Distillation Techniques Nov 22, 2024 Hate Speech Detection Knowledge Distillation
— Unverified 00 BAPO: Base-Anchored Preference Optimization for Overcoming Forgetting in Large Language Models Personalization Jun 30, 2024 Continual Learning General Knowledge
— Unverified 00 Bayesian-Optimized One-Step Diffusion Model with Knowledge Distillation for Real-Time 3D Human Motion Prediction Sep 19, 2024 Bayesian Optimization Human motion prediction
— Unverified 00 BD-KD: Balancing the Divergences for Online Knowledge Distillation Dec 25, 2022 Knowledge Distillation Model Compression
— Unverified 00