LTD: Low Temperature Distillation for Robust Adversarial Training Nov 3, 2021 Knowledge Distillation
— Unverified 0Leveraging Advantages of Interactive and Non-Interactive Models for Vector-Based Cross-Lingual Information Retrieval Nov 3, 2021 Computational Efficiency Cross-Lingual Information Retrieval
— Unverified 0Knowledge Cross-Distillation for Membership Privacy Nov 2, 2021 Inference Attack Knowledge Distillation
— Unverified 0Domain-Lifelong Learning for Dialogue State Tracking via Knowledge Preservation Networks Nov 1, 2021 Dialogue State Tracking Diversity
Code Code Available 0Universal-KD: Attention-based Output-Grounded Intermediate Layer Knowledge Distillation Nov 1, 2021 Knowledge Distillation
— Unverified 0deepQuest-py: Large and Distilled Models for Quality Estimation Nov 1, 2021 Knowledge Distillation Sentence
Code Code Available 0Papago’s Submission for the WMT21 Quality Estimation Shared Task Nov 1, 2021 Knowledge Distillation Multi-Task Learning
— Unverified 0HW-TSC’s Participation in the WMT 2021 Large-Scale Multilingual Translation Task Nov 1, 2021 Knowledge Distillation Translation
— Unverified 0HW-TSC’s Participation in the WMT 2021 News Translation Shared Task Nov 1, 2021 de-en Knowledge Distillation
— Unverified 0Students Who Study Together Learn Better: On the Importance of Collective Knowledge Distillation for Domain Transfer in Fact Verification Nov 1, 2021 Fact Verification Knowledge Distillation
— Unverified 0Limitations of Knowledge Distillation for Zero-shot Transfer Learning Nov 1, 2021 CPU Cross-Lingual Transfer
— Unverified 0The NiuTrans System for the WMT 2021 Efficiency Task Nov 1, 2021 GPU Knowledge Distillation
— Unverified 0How to Select One Among All ? An Empirical Study Towards the Robustness of Knowledge Distillation in Natural Language Understanding Nov 1, 2021 Adversarial Robustness All
— Unverified 0PDALN: Progressive Domain Adaptation over a Pre-trained Model for Low-Resource Cross-Domain Named Entity Recognition Nov 1, 2021 Cross-Domain Named Entity Recognition Data Augmentation
— Unverified 0Mutual-Learning Improves End-to-End Speech Translation Nov 1, 2021 Knowledge Distillation Machine Translation
— Unverified 0The Mininglamp Machine Translation System for WMT21 Nov 1, 2021 Knowledge Distillation Machine Translation
— Unverified 0Exploring Non-Autoregressive Text Style Transfer Nov 1, 2021 Contrastive Learning Knowledge Distillation
Code Code Available 0The LMU Munich System for the WMT 2021 Large-Scale Multilingual Machine Translation Shared Task Nov 1, 2021 Data Augmentation Knowledge Distillation
— Unverified 0Multilingual Neural Machine Translation: Can Linguistic Hierarchies Help? Nov 1, 2021 Knowledge Distillation Machine Translation
— Unverified 0AUTOSUMM: Automatic Model Creation for Text Summarization Nov 1, 2021 Abstractive Text Summarization Deep Learning
— Unverified 0RW-KD: Sample-wise Loss Terms Re-Weighting for Knowledge Distillation Nov 1, 2021 Knowledge Distillation
— Unverified 0TenTrans Large-Scale Multilingual Machine Translation System for WMT21 Nov 1, 2021 Knowledge Distillation Machine Translation
— Unverified 0GAML-BERT: Improving BERT Early Exiting by Gradient Aligned Mutual Learning Nov 1, 2021 Knowledge Distillation
— Unverified 0Improving Stance Detection with Multi-Dataset Learning and Knowledge Distillation Nov 1, 2021 Knowledge Distillation Stance Detection
Code Code Available 0Efficient Machine Translation with Model Pruning and Quantization Nov 1, 2021 CPU Decoder
— Unverified 0NVIDIA NeMo’s Neural Machine Translation Systems for English-German and English-Russian News and Biomedical Tasks at WMT21 Nov 1, 2021 Data Augmentation Knowledge Distillation
— Unverified 0PP-ShiTu: A Practical Lightweight Image Recognition System Nov 1, 2021 Face Recognition Knowledge Distillation
Code Code Available 0Distilling Knowledge for Empathy Detection Nov 1, 2021 Knowledge Distillation
Code Code Available 0Collaborative Learning of Bidirectional Decoders for Unsupervised Text Style Transfer Nov 1, 2021 Attribute Decoder
Code Code Available 0Combining Curriculum Learning and Knowledge Distillation for Dialogue Generation Nov 1, 2021 Dialogue Generation Knowledge Distillation
— Unverified 0Rethinking the Knowledge Distillation From the Perspective of Model Calibration Oct 31, 2021 Knowledge Distillation
— Unverified 0Estimating and Maximizing Mutual Information for Knowledge Distillation Oct 29, 2021 Knowledge Distillation
— Unverified 0On Cross-Layer Alignment for Model Fusion of Heterogeneous Neural Networks Oct 29, 2021 Knowledge Distillation Model Compression
— Unverified 0Towards Model Agnostic Federated Learning Using Knowledge Distillation Oct 28, 2021 Federated Learning Knowledge Distillation
— Unverified 0NxMTransformer: Semi-Structured Sparsification for Natural Language Understanding via ADMM Oct 28, 2021 Knowledge Distillation Natural Language Understanding
— Unverified 0Temporal Knowledge Distillation for On-device Audio Classification Oct 27, 2021 Audio Classification Classification
— Unverified 0GenURL: A General Framework for Unsupervised Representation Learning Oct 27, 2021 Contrastive Learning Dimensionality Reduction
— Unverified 0Beyond Classification: Knowledge Distillation using Multi-Object Impressions Oct 27, 2021 Classification Knowledge Distillation
— Unverified 0Response-based Distillation for Incremental Object Detection Oct 26, 2021 Incremental Learning Knowledge Distillation
— Unverified 0MUSE: Feature Self-Distillation with Mutual Information and Self-Information Oct 25, 2021 image-classification Image Classification
— Unverified 0Reconstructing Pruned Filters using Cheap Spatial Transformations Oct 25, 2021 Feature Compression Knowledge Distillation
— Unverified 0X-Distill: Improving Self-Supervised Monocular Depth via Cross-Task Distillation Oct 24, 2021 Depth Estimation Knowledge Distillation
— Unverified 0Pseudo Supervised Monocular Depth Estimation with Teacher-Student Network Oct 22, 2021 Depth Estimation Knowledge Distillation
— Unverified 0How and When Adversarial Robustness Transfers in Knowledge Distillation? Oct 22, 2021 Adversarial Robustness Knowledge Distillation
— Unverified 0Augmenting Knowledge Distillation With Peer-To-Peer Mutual Learning For Model Compression Oct 21, 2021 Knowledge Distillation Model Compression
— Unverified 0Class Incremental Online Streaming Learning Oct 20, 2021 class-incremental learning Class Incremental Learning
— Unverified 0Knowledge distillation from language model to acoustic model: a hierarchical multi-task learning approach Oct 20, 2021 Knowledge Distillation Language Modeling
— Unverified 0FedHe: Heterogeneous Models and Communication-Efficient Federated Learning Oct 19, 2021 Federated Learning Knowledge Distillation
Code Code Available 0Adaptive Distillation: Aggregating Knowledge from Multiple Paths for Efficient Distillation Oct 19, 2021 Knowledge Distillation Neural Network Compression
Code Code Available 0HRKD: Hierarchical Relational Knowledge Distillation for Cross-domain Language Model Compression Oct 16, 2021 Few-Shot Learning Knowledge Distillation
Code Code Available 0