MiniDisc: Minimal Distillation Schedule for Language Model Compression May 29, 2022 Knowledge Distillation Language Modeling
Code Code Available 0Divide to Adapt: Mitigating Confirmation Bias for Domain Adaptation of Black-Box Predictors May 28, 2022 Domain Adaptation Knowledge Distillation
Code Code Available 1One Reference Is Not Enough: Diverse Distillation with Reference Selection for Non-Autoregressive Translation May 28, 2022 Knowledge Distillation Machine Translation
Code Code Available 0Parameter-Efficient and Student-Friendly Knowledge Distillation May 28, 2022 Knowledge Distillation Transfer Learning
— Unverified 0Geometer: Graph Few-Shot Class-Incremental Learning via Prototype Representation May 27, 2022 class-incremental learning Class Incremental Learning
Code Code Available 1Continual evaluation for lifelong learning: Identifying the stability gap May 26, 2022 Continual Learning Incremental Learning
Code Code Available 1Region-aware Knowledge Distillation for Efficient Image-to-Image Translation May 25, 2022 Contrastive Learning image-classification
— Unverified 0Do we need Label Regularization to Fine-tune Pre-trained Language Models? May 25, 2022 Knowledge Distillation Model Compression
— Unverified 0DFM: Dialogue Foundation Model for Universal Large-Scale Dialogue-Oriented Task Learning May 25, 2022 Dialogue Generation Diversity
— Unverified 0Compressing Deep Graph Neural Networks via Adversarial Knowledge Distillation May 24, 2022 Graph Classification Knowledge Distillation
Code Code Available 1Optimizing Performance of Federated Person Re-identification: Benchmarking and Analysis May 24, 2022 Benchmarking Federated Learning
Code Code Available 1CDFKD-MFS: Collaborative Data-free Knowledge Distillation via Multi-level Feature Sharing May 24, 2022 Data-free Knowledge Distillation Knowledge Distillation
Code Code Available 0IDEAL: Query-Efficient Data-Free Learning from Black-box Models May 23, 2022 Knowledge Distillation
Code Code Available 1Boosting Multi-Label Image Classification with Complementary Parallel Self-Distillation May 23, 2022 image-classification Image Classification
Code Code Available 1PointDistiller: Structured Knowledge Distillation Towards Efficient and Compact 3D Detection May 23, 2022 3D Object Detection Knowledge Distillation
Code Code Available 1LILA-BOTI : Leveraging Isolated Letter Accumulations By Ordering Teacher Insights for Bangla Handwriting Recognition May 23, 2022 Handwriting Recognition Knowledge Distillation
Code Code Available 0Knowledge Distillation via the Target-aware Transformer May 22, 2022 Knowledge Distillation
Code Code Available 1Aligning Logits Generatively for Principled Black-Box Knowledge Distillation May 21, 2022 Federated Learning Knowledge Distillation
Code Code Available 0Knowledge Distillation from A Stronger Teacher May 21, 2022 image-classification Image Classification
Code Code Available 1Exploring Extreme Parameter Compression for Pre-trained Language Models May 20, 2022 Knowledge Distillation Tensor Decomposition
Code Code Available 1InDistill: Information flow-preserving knowledge distillation for model compression May 20, 2022 Knowledge Distillation Model Compression
Code Code Available 0Simple Regularisation for Uncertainty-Aware Knowledge Distillation May 19, 2022 BIG-bench Machine Learning Diversity
— Unverified 0ERNIE-Search: Bridging Cross-Encoder with Dual-Encoder via Self On-the-fly Distillation for Dense Passage Retrieval May 18, 2022 Knowledge Distillation Open-Domain Question Answering
— Unverified 0Prompting to Distill: Boosting Data-Free Knowledge Distillation via Reinforced Prompt May 16, 2022 Data-free Knowledge Distillation Knowledge Distillation
— Unverified 0Chemical transformer compression for accelerating both training and inference of molecular modeling May 16, 2022 Knowledge Distillation Model Compression
Code Code Available 0Directed Acyclic Transformer for Non-Autoregressive Machine Translation May 16, 2022 Knowledge Distillation Machine Translation
Code Code Available 1Not to Overfit or Underfit the Source Domains? An Empirical Study of Domain Generalization in Question Answering May 15, 2022 Domain Generalization Knowledge Distillation
— Unverified 0Knowledge Distillation Meets Open-Set Semi-Supervised Learning May 13, 2022 Face Recognition Knowledge Distillation
Code Code Available 1"Teaching Independent Parts Separately" (TIPSy-GAN) : Improving Accuracy and Stability in Unsupervised Adversarial 2D to 3D Pose Estimation May 12, 2022 3D Human Pose Estimation 3D Pose Estimation
— Unverified 0D3T-GAN: Data-Dependent Domain Transfer GANs for Few-shot Image Generation May 12, 2022 Image Generation Knowledge Distillation
— Unverified 0Knowledge Distillation for Multi-Target Domain Adaptation in Real-Time Person Re-Identification May 12, 2022 Domain Adaptation Knowledge Distillation
Code Code Available 0DistilProtBert: A distilled protein language model used to distinguish between real proteins and their randomly shuffled counterparts May 10, 2022 Dimensionality Reduction Knowledge Distillation
Code Code Available 1Incremental-DETR: Incremental Few-Shot Object Detection via Self-Supervised Learning May 9, 2022 Few-Shot Object Detection Knowledge Distillation
— Unverified 0Data-Free Adversarial Knowledge Distillation for Graph Neural Networks May 8, 2022 Generative Adversarial Network Graph Classification
— Unverified 0ConceptDistil: Model-Agnostic Distillation of Concept Explanations May 7, 2022 Explainable Models Knowledge Distillation
— Unverified 0Automatic Block-wise Pruning with Auxiliary Gating Structures for Deep Convolutional Neural Networks May 7, 2022 Knowledge Distillation Model Compression
— Unverified 0Distilling Inter-Class Distance for Semantic Segmentation May 7, 2022 Knowledge Distillation Position
— Unverified 0Collective Relevance Labeling for Passage Retrieval May 6, 2022 Information Retrieval Knowledge Distillation
Code Code Available 0Alignahead: Online Cross-Layer Knowledge Extraction on Graph Neural Networks May 5, 2022 Knowledge Distillation
Code Code Available 0Holistic Approach to Measure Sample-level Adversarial Vulnerability and its Utility in Building Trustworthy Systems May 5, 2022 Adversarial Attack Knowledge Distillation
— Unverified 0A Deep Reinforcement Learning Framework for Rapid Diagnosis of Whole Slide Pathological Images May 5, 2022 Deep Reinforcement Learning Knowledge Distillation
— Unverified 0Spot-adaptive Knowledge Distillation May 5, 2022 Knowledge Distillation
Code Code Available 1FedSPLIT: One-Shot Federated Recommendation System Based on Non-negative Joint Matrix Factorization and Knowledge Distillation May 4, 2022 Collaborative Filtering Federated Learning
— Unverified 0Attention-based Knowledge Distillation in Multi-attention Tasks: The Impact of a DCT-driven Loss May 4, 2022 Descriptive Knowledge Distillation
— Unverified 0Knowledge Distillation of Russian Language Models with Reduction of Vocabulary May 4, 2022 Knowledge Distillation
Code Code Available 0Generalized Knowledge Distillation via Relationship Matching May 4, 2022 Few-Shot Learning Incremental Learning
Code Code Available 0Masked Generative Distillation May 3, 2022 image-classification Image Classification
Code Code Available 2FedDKD: Federated Learning with Decentralized Knowledge Distillation May 2, 2022 Federated Learning Knowledge Distillation
— Unverified 0Redistributing Low-Frequency Words: Making the Most of Monolingual Data in Non-Autoregressive Translation May 1, 2022 Knowledge Distillation Translation
Code Code Available 0Knowledge Distillation Meets Few-Shot Learning: An Approach for Few-Shot Intent Classification Within and Across Domains May 1, 2022 Cross-Domain Few-Shot cross-domain few-shot learning
— Unverified 0