Understanding Knowledge Distillation in Non-autoregressive Machine Translation Nov 7, 2019 Knowledge Distillation Machine Translation
— Unverified 0Data Diversification: A Simple Strategy For Neural Machine Translation Nov 5, 2019 Knowledge Distillation Machine Translation
Code Code Available 1ESPnet How2 Speech Translation System for IWSLT 2019: Pre-training, Knowledge Distillation, and Going Deeper Nov 1, 2019 All Knowledge Distillation
— Unverified 0Weakly Supervised Cross-lingual Semantic Relation Classification via Knowledge Distillation Nov 1, 2019 Classification Cross-Lingual Transfer
— Unverified 0Natural Language Generation for Effective Knowledge Distillation Nov 1, 2019 Knowledge Distillation Linguistic Acceptability
Code Code Available 0Distilling Pixel-Wise Feature Similarities for Semantic Segmentation Oct 31, 2019 Knowledge Distillation Neural Network Compression
— Unverified 0A Simple but Effective BERT Model for Dialog State Tracking on Resource-Limited Systems Oct 28, 2019 dialog state tracking Dialogue State Tracking
— Unverified 0MOD: A Deep Mixture Model with Online Knowledge Distillation for Large Scale Video Temporal Concept Localization Oct 27, 2019 Knowledge Distillation Video Understanding
Code Code Available 0Variational Student: Learning Compact and Sparser Networks in Knowledge Distillation Framework Oct 26, 2019 Knowledge Distillation Variational Inference
— Unverified 0Secost: Sequential co-supervision for large scale weakly labeled audio event detection Oct 25, 2019 Event Detection Knowledge Distillation
— Unverified 0An Empirical Study of Efficient ASR Rescoring with Transformers Oct 24, 2019 Knowledge Distillation Language Modeling
— Unverified 0Adversarial Feature Alignment: Avoid Catastrophic Forgetting in Incremental Task Lifelong Learning Oct 24, 2019 Continual Learning image-classification
— Unverified 0Contrastive Representation Distillation Oct 23, 2019 Contrastive Learning Knowledge Distillation
Code Code Available 1Model Compression with Two-stage Multi-teacher Knowledge Distillation for Web Question Answering System Oct 18, 2019 General Knowledge Knowledge Distillation
— Unverified 0A Generalized and Robust Method Towards Practical Gaze Estimation on Smart Phone Oct 16, 2019 Gaze Estimation Knowledge Distillation
— Unverified 0Noise as a Resource for Learning in Knowledge Distillation Oct 11, 2019 Knowledge Distillation
— Unverified 0VarGFaceNet: An Efficient Variable Group Convolutional Neural Network for Lightweight Face Recognition Oct 11, 2019 Face Detection Face Identification
Code Code Available 0Cross-modal knowledge distillation for action recognition Oct 10, 2019 Action Recognition Knowledge Distillation
— Unverified 0FedMD: Heterogenous Federated Learning via Model Distillation Oct 8, 2019 Federated Learning Knowledge Distillation
Code Code Available 1Knowledge Distillation from Internal Representations Oct 8, 2019 Knowledge Distillation
— Unverified 0Distilling BERT into Simple Neural Networks with Unlabeled Transfer Data Oct 4, 2019 Knowledge Distillation NER
— Unverified 0On the Efficacy of Knowledge Distillation Oct 3, 2019 Knowledge Distillation
— Unverified 0Improving Word Embedding Factorization for Compression Using Distilled Nonlinear Neural Decomposition Oct 2, 2019 Knowledge Distillation Language Modeling
— Unverified 0AntMan: Sparse Low-Rank Compression to Accelerate RNN inference Oct 2, 2019 Knowledge Distillation Low-rank compression
— Unverified 0DistilBERT, a distilled version of BERT: smaller, faster, cheaper and lighter Oct 2, 2019 Hate Speech Detection Knowledge Distillation
Code Code Available 1Distilled Split Deep Neural Networks for Edge-Assisted Real-Time Systems Oct 1, 2019 Edge-computing Image Classification
Code Code Available 1A Bayesian Optimization Framework for Neural Network Compression Oct 1, 2019 Bayesian Optimization Knowledge Distillation
— Unverified 0Distillation-Based Training for Multi-Exit Architectures Oct 1, 2019 Knowledge Distillation
Code Code Available 1Training convolutional neural networks with cheap convolutions and online distillation Sep 28, 2019 Knowledge Distillation
Code Code Available 0Compact Trilinear Interaction for Visual Question Answering Sep 26, 2019 Benchmarking Knowledge Distillation
Code Code Available 0Distilled embedding: non-linear embedding factorization using knowledge distillation Sep 25, 2019 Knowledge Distillation Machine Translation
— Unverified 0Collaborative Inter-agent Knowledge Distillation for Reinforcement Learning Sep 25, 2019 Decision Making Knowledge Distillation
— Unverified 0Proactive Sequence Generator via Knowledge Acquisition Sep 25, 2019 de-en Knowledge Distillation
— Unverified 0XD: Cross-lingual Knowledge Distillation for Polyglot Sentence Embeddings Sep 25, 2019 Knowledge Distillation Language Modeling
— Unverified 0SELF-KNOWLEDGE DISTILLATION ADVERSARIAL ATTACK Sep 25, 2019 Adversarial Attack Knowledge Distillation
— Unverified 0Revisiting Knowledge Distillation via Label Smoothing Regularization Sep 25, 2019 Knowledge Distillation Self-Knowledge Distillation
Code Code Available 0Extremely Small BERT Models from Mixed-Vocabulary Training Sep 25, 2019 Knowledge Distillation Language Modelling
— Unverified 0Technical report on Conversational Question Answering Sep 24, 2019 Conversational Question Answering Data Augmentation
— Unverified 0FEED: Feature-level Ensemble for Knowledge Distillation Sep 24, 2019 Knowledge Distillation
— Unverified 0TinyBERT: Distilling BERT for Natural Language Understanding Sep 23, 2019 Knowledge Distillation Language Modelling
Code Code Available 0Positive-Unlabeled Compression on the Cloud Sep 21, 2019 GPU Knowledge Distillation
Code Code Available 2Learning Lightweight Pedestrian Detector with Hierarchical Knowledge Distillation Sep 20, 2019 Knowledge Distillation Pedestrian Detection
— Unverified 0Ensemble Knowledge Distillation for Learning Improved and Efficient Networks Sep 17, 2019 Ensemble Learning General Classification
Code Code Available 0Knowledge Transfer Graph for Deep Collaborative Learning Sep 10, 2019 Knowledge Distillation Transfer Learning
Code Code Available 0Accelerating Transformer Decoding via a Hybrid of Self-attention and Recurrent Neural Network Sep 5, 2019 Decoder Knowledge Distillation
— Unverified 0Knowledge distillation for optimization of quantized deep neural networks Sep 4, 2019 Knowledge Distillation
— Unverified 0Knowledge Distillation for End-to-End Person Search Sep 3, 2019 Knowledge Distillation Model Compression
Code Code Available 0Online Sensor Hallucination via Knowledge Distillation for Multimodal Image Classification Aug 28, 2019 Classification Decision Making
— Unverified 0Patient Knowledge Distillation for BERT Model Compression Aug 25, 2019 Knowledge Distillation model
Code Code Available 0Well-Read Students Learn Better: On the Importance of Pre-training Compact Models Aug 23, 2019 Knowledge Distillation Language Modelling
Code Code Available 2