Adaptive Multi-Teacher Knowledge Distillation with Meta-Learning Jun 11, 2023 Knowledge Distillation Meta-Learning
Code Code Available 1GKD: A General Knowledge Distillation Framework for Large-scale Pre-trained Language Model Jun 11, 2023 General Knowledge Knowledge Distillation
Code Code Available 1EaSyGuide : ESG Issue Identification Framework leveraging Abilities of Generative Large Language Models Jun 11, 2023 Articles Knowledge Distillation
Code Code Available 0Are Intermediate Layers and Labels Really Necessary? A General Language Model Distillation Method Jun 11, 2023 Knowledge Distillation Language Modeling
Code Code Available 1Improving Frame-level Classifier for Word Timings with Non-peaky CTC in End-to-End Automatic Speech Recognition Jun 9, 2023 Automatic Speech Recognition Automatic Speech Recognition (ASR)
— Unverified 0RankFormer: Listwise Learning-to-Rank Using Listwide Labels Jun 9, 2023 Knowledge Distillation Learning-To-Rank
Code Code Available 1The economic trade-offs of large language models: A case study Jun 8, 2023 Knowledge Distillation Prompt Engineering
— Unverified 0BOOT: Data-free Distillation of Denoising Diffusion Models with Bootstrapping Jun 8, 2023 Denoising Knowledge Distillation
— Unverified 0Population-Based Evolutionary Gaming for Unsupervised Person Re-identification Jun 8, 2023 Diversity Knowledge Distillation
— Unverified 0Self-supervised Audio Teacher-Student Transformer for Both Clip-level and Frame-level Tasks Jun 7, 2023 Audio Classification Audio Tagging
Code Code Available 1Faithful Knowledge Distillation Jun 7, 2023 Adversarial Robustness Knowledge Distillation
— Unverified 0Model-Based Reinforcement Learning with Multi-Task Offline Pretraining Jun 6, 2023 Knowledge Distillation Model-based Reinforcement Learning
Code Code Available 0Orca: Progressive Learning from Complex Explanation Traces of GPT-4 Jun 5, 2023 Imitation Learning Knowledge Distillation
Code Code Available 1Zero shot framework for satellite image restoration Jun 5, 2023 Disentanglement Image Restoration
— Unverified 0Joint Pre-training and Local Re-training: Transferable Representation Learning on Multi-source Knowledge Graphs Jun 5, 2023 Entity Alignment Knowledge Distillation
Code Code Available 0I^3 Retriever: Incorporating Implicit Interaction in Pre-trained Language Models for Passage Retrieval Jun 4, 2023 Knowledge Distillation Passage Retrieval
Code Code Available 1Revisiting Data-Free Knowledge Distillation with Poisoned Teachers Jun 4, 2023 Backdoor Defense for Data-Free Distillation with Poisoned Teachers Data-free Knowledge Distillation
Code Code Available 1Modular Transformers: Compressing Transformers into Modularized Layers for Flexible Efficient Inference Jun 4, 2023 Decoder Knowledge Distillation
— Unverified 0Evolving Knowledge Mining for Class Incremental Segmentation Jun 3, 2023 Class-Incremental Semantic Segmentation Knowledge Distillation
Code Code Available 0Deep Classifier Mimicry without Data Access Jun 3, 2023 Knowledge Distillation
Code Code Available 0Group channel pruning and spatial attention distilling for object detection Jun 2, 2023 Knowledge Distillation Model Compression
— Unverified 0Privacy Distillation: Reducing Re-identification Risk of Multimodal Diffusion Models Jun 2, 2023 Knowledge Distillation
— Unverified 0Speech Translation with Foundation Models and Optimal Transport: UPC at IWSLT23 Jun 2, 2023 Knowledge Distillation Machine Translation
— Unverified 0Improved Cross-Lingual Transfer Learning For Automatic Speech Translation Jun 1, 2023 automatic-speech-translation Cross-Lingual Transfer
— Unverified 0Teacher Agent: A Knowledge Distillation-Free Framework for Rehearsal-based Video Incremental Learning Jun 1, 2023 Incremental Learning Knowledge Distillation
Code Code Available 0Accurate and Structured Pruning for Efficient Automatic Speech Recognition May 31, 2023 Automatic Speech Recognition Automatic Speech Recognition (ASR)
— Unverified 0Graph Entropy Minimization for Semi-supervised Node Classification May 31, 2023 Classification Knowledge Distillation
Code Code Available 0PlaSma: Making Small Language Models Better Procedural Knowledge Models for (Counterfactual) Planning May 31, 2023 Common Sense Reasoning counterfactual
Code Code Available 1KEYword based Sampling (KEYS) for Large Language Models May 30, 2023 Knowledge Distillation Language Modeling
— Unverified 0Are Large Kernels Better Teachers than Transformers for ConvNets? May 30, 2023 Knowledge Distillation
Code Code Available 2Research on Multilingual News Clustering Based on Cross-Language Word Embeddings May 30, 2023 Clustering Knowledge Distillation
— Unverified 0Semi-supervised Pathological Image Segmentation via Cross Distillation of Multiple Attentions May 30, 2023 Decoder Image Segmentation
Code Code Available 1A Recipe for Efficient SBIR Models: Combining Relative Triplet Loss with Batch Normalization and Knowledge Distillation May 30, 2023 Data Augmentation Image Retrieval
— Unverified 0Bridging the Sim-to-Real Gap from the Information Bottleneck Perspective May 29, 2023 Knowledge Distillation Reinforcement Learning (RL)
Code Code Available 0GripRank: Bridging the Gap between Retrieval and Generation via the Generative Knowledge Improved Passage Ranking May 29, 2023 Answer Generation Dialogue Generation
— Unverified 0Learning to Learn from APIs: Black-Box Data-Free Meta-Learning May 28, 2023 Few-Shot Learning Knowledge Distillation
Code Code Available 1DPHuBERT: Joint Distillation and Pruning of Self-Supervised Speech Models May 28, 2023 Knowledge Distillation Self-Supervised Learning
Code Code Available 1ConaCLIP: Exploring Distillation of Fully-Connected Knowledge Interaction Graph for Lightweight Text-Image Retrieval May 28, 2023 Image Retrieval Knowledge Distillation
— Unverified 0Towards Better Entity Linking with Multi-View Enhanced Distillation May 27, 2023 Entity Linking Knowledge Distillation
Code Code Available 1FoPro-KD: Fourier Prompted Effective Knowledge Distillation for Long-Tailed Medical Image Recognition May 27, 2023 image-classification Image Classification
Code Code Available 1One-Step Knowledge Distillation and Fine-Tuning in Using Large Pre-Trained Self-Supervised Learning Models for Speaker Verification May 27, 2023 Knowledge Distillation Self-Supervised Learning
Code Code Available 1Knowledge Distillation Performs Partial Variance Reduction May 27, 2023 Knowledge Distillation
Code Code Available 0Vision Transformers for Small Histological Datasets Learned through Knowledge Distillation May 27, 2023 Airbubbles Detection Anomaly Detection
Code Code Available 0Improving Knowledge Distillation via Regularizing Feature Norm and Direction May 26, 2023 Domain Adaptation Knowledge Distillation
Code Code Available 1ABC-KD: Attention-Based-Compression Knowledge Distillation for Deep Learning-Based Noise Suppression May 26, 2023 Knowledge Distillation
— Unverified 0A Study on Knowledge Distillation from Weak Teacher for Scaling Up Pre-trained Language Models May 26, 2023 Knowledge Distillation
— Unverified 0Knowledge Diffusion for Distillation May 25, 2023 Denoising image-classification
Code Code Available 1Cross-Lingual Knowledge Distillation for Answer Sentence Selection in Low-Resource Languages May 25, 2023 Knowledge Distillation Machine Translation
— Unverified 0OVO: Open-Vocabulary Occupancy May 25, 2023 Knowledge Distillation Prediction
Code Code Available 1On the Impact of Knowledge Distillation for Model Interpretability May 25, 2023 Knowledge Distillation
— Unverified 0