Cross-Lingual Knowledge Distillation for Answer Sentence Selection in Low-Resource Languages May 25, 2023 Knowledge Distillation Machine Translation
— Unverified 0OVO: Open-Vocabulary Occupancy May 25, 2023 Knowledge Distillation Prediction
Code Code Available 1Fairness Continual Learning Approach to Semantic Scene Understanding in Open-World Environments May 25, 2023 Continual Learning Continual Semantic Segmentation
— Unverified 0Collective Knowledge Graph Completion with Mutual Knowledge Distillation May 25, 2023 Knowledge Distillation Knowledge Graph Completion
— Unverified 0Knowledge Distillation for Neural Transducer-based Target-Speaker ASR: Exploiting Parallel Mixture/Single-Talker Speech Data May 25, 2023 Knowledge Distillation Speech Extraction
— Unverified 0Triplet Knowledge Distillation May 25, 2023 Face Recognition image-classification
— Unverified 0Camera-Incremental Object Re-Identification with Identity Knowledge Evolution May 25, 2023 Knowledge Distillation Object
Code Code Available 0How to Distill your BERT: An Empirical Study on the Impact of Weight Initialisation and Distillation Objectives May 24, 2023 Knowledge Distillation QNLI
Code Code Available 1HARD: Hard Augmentations for Robust Distillation May 24, 2023 Data Augmentation Domain Generalization
— Unverified 0Incorporating Ultrasound Tongue Images for Audio-Visual Speech Enhancement through Knowledge Distillation May 24, 2023 Automatic Speech Recognition Automatic Speech Recognition (ASR)
— Unverified 0CoLaDa: A Collaborative Label Denoising Framework for Cross-lingual Named Entity Recognition May 24, 2023 Denoising Knowledge Distillation
Code Code Available 2Deakin RF-Sensing: Experiments on Correlated Knowledge Distillation for Monitoring Human Postures with Radios May 24, 2023 Knowledge Distillation
— Unverified 0Just CHOP: Embarrassingly Simple LLM Compression May 24, 2023 Knowledge Distillation Language Modeling
— Unverified 0AdvFunMatch: When Consistent Teaching Meets Adversarial Robustness May 24, 2023 Adversarial Robustness Knowledge Distillation
— Unverified 0PruMUX: Augmenting Data Multiplexing with Model Compression May 24, 2023 Knowledge Distillation model
Code Code Available 0NORM: Knowledge Distillation via N-to-One Representation Matching May 23, 2023 Knowledge Distillation
Code Code Available 1Sequence-Level Knowledge Distillation for Class-Incremental End-to-End Spoken Language Understanding May 23, 2023 Continual Learning Decoder
— Unverified 0One-stop Training of Multiple Capacity Models May 23, 2023 Knowledge Distillation Machine Translation
— Unverified 0Masked Modeling Duo for Speech: Specializing General-Purpose Audio Representation to Speech using Denoising Distillation May 23, 2023 Denoising Knowledge Distillation
— Unverified 0Decoupled Kullback-Leibler Divergence Loss May 23, 2023 Adversarial Defense Adversarial Robustness
Code Code Available 1Transferring Learning Trajectories of Neural Networks May 23, 2023 Knowledge Distillation
— Unverified 0EnSiam: Self-Supervised Learning With Ensemble Representations May 22, 2023 Contrastive Learning Knowledge Distillation
— Unverified 0Distilling Robustness into Natural Language Inference Models with Domain-Targeted Augmentation May 22, 2023 Data Augmentation Knowledge Distillation
— Unverified 0Lion: Adversarial Distillation of Proprietary Large Language Models May 22, 2023 Instruction Following Knowledge Distillation
Code Code Available 2Is Synthetic Data From Diffusion Models Ready for Knowledge Distillation? May 22, 2023 Data-free Knowledge Distillation Few-Shot Learning
Code Code Available 1D^2TV: Dual Knowledge Distillation and Target-oriented Vision Modeling for Many-to-Many Multimodal Summarization May 22, 2023 Knowledge Distillation
Code Code Available 0Revisiting Data Augmentation in Model Compression: An Empirical and Comprehensive Study May 22, 2023 Data Augmentation Knowledge Distillation
— Unverified 0Understanding the Effect of Data Augmentation on Knowledge Distillation May 21, 2023 Data Augmentation Knowledge Distillation
— Unverified 0DualVC: Dual-mode Voice Conversion using Intra-model Knowledge Distillation and Hybrid Predictive Coding May 21, 2023 Data Augmentation Decoder
— Unverified 0One-Shot Federated Learning for LEO Constellations that Reduces Convergence Time from Days to 90 Minutes May 21, 2023 Federated Learning Knowledge Distillation
— Unverified 0DisCo: Distilled Student Models Co-training for Semi-supervised Text Mining May 20, 2023 Extractive Summarization Knowledge Distillation
Code Code Available 1Lifting the Curse of Capacity Gap in Distilling Language Models May 20, 2023 Knowledge Distillation Mixture-of-Experts
Code Code Available 1Accurate Knowledge Distillation with n-best Reranking May 20, 2023 Knowledge Distillation Reranking
— Unverified 0Sentence Embedder Guided Utterance Encoder (SEGUE) for Spoken Language Understanding May 20, 2023 Knowledge Distillation Sentence
Code Code Available 0Pseudo-Label Training and Model Inertia in Neural Machine Translation May 19, 2023 Knowledge Distillation Machine Translation
— Unverified 0Boost Vision Transformer with GPU-Friendly Sparsity and Quantization May 18, 2023 Benchmarking GPU
— Unverified 0BERM: Training the Balanced and Extractable Representation for Matching to Improve Generalization Ability of Dense Retrieval May 18, 2023 Information Retrieval Knowledge Distillation
— Unverified 0Cross-modality Data Augmentation for End-to-End Sign Language Translation May 18, 2023 Data Augmentation Knowledge Distillation
Code Code Available 1DQ-Whisper: Joint Distillation and Quantization for Efficient Multilingual Speech Recognition May 18, 2023 Knowledge Distillation Quantization
— Unverified 0Student-friendly Knowledge Distillation May 18, 2023 Knowledge Distillation
— Unverified 0Catch-Up Distillation: You Only Need to Train Once for Accelerating Sampling May 18, 2023 Knowledge Distillation
Code Code Available 0AD-KD: Attribution-Driven Knowledge Distillation for Language Model Compression May 17, 2023 Knowledge Distillation Language Modeling
Code Code Available 1When Gradient Descent Meets Derivative-Free Optimization: A Match Made in Black-Box Scenario May 17, 2023 Knowledge Distillation
— Unverified 0Tailoring Instructions to Student's Learning Levels Boosts Knowledge Distillation May 16, 2023 Knowledge Distillation text-classification
Code Code Available 1Weight-Inherited Distillation for Task-Agnostic BERT Compression May 16, 2023 Knowledge Distillation
Code Code Available 0Lightweight Self-Knowledge Distillation with Multi-source Information Fusion May 16, 2023 Knowledge Distillation Self-Knowledge Distillation
Code Code Available 0Bridging the Domain Gap: Self-Supervised 3D Scene Understanding with Foundation Models May 15, 2023 3D Object Detection Image Captioning
Code Code Available 1Soft Prompt Decoding for Multilingual Dense Retrieval May 15, 2023 Cross-Lingual Information Retrieval Information Retrieval
— Unverified 0Distilling Knowledge for Short-to-Long Term Trajectory Prediction May 15, 2023 Knowledge Distillation Prediction
— Unverified 0Improving Defensive Distillation using Teacher Assistant May 14, 2023 Face Recognition Knowledge Distillation
— Unverified 0