Technical Report for ICCV 2021 Challenge SSLAD-Track3B: Transformers Are Better Continual Learners Jan 13, 2022 Continual Learning Knowledge Distillation
— Unverified 00 Technical Report of Team GraphMIRAcles in the WikiKG90M-LSC Track of OGB-LSC @ KDD Cup 2021 Jul 12, 2021 Knowledge Distillation Knowledge Graphs
— Unverified 00 Technical report on Conversational Question Answering Sep 24, 2019 Conversational Question Answering Data Augmentation
— Unverified 00 Temporal Knowledge Distillation for On-device Audio Classification Oct 27, 2021 Audio Classification Classification
— Unverified 00 Temporal Knowledge Distillation for Time-Sensitive Financial Services Applications Dec 28, 2023 Anomaly Detection Fraud Detection
— Unverified 00 Temporal reasoning for timeline summarisation in social media Dec 30, 2024 Knowledge Distillation Timeline Summarization
— Unverified 00 Temporal Separation with Entropy Regularization for Knowledge Distillation in Spiking Neural Networks Mar 5, 2025 Computational Efficiency Knowledge Distillation
— Unverified 00 TenTrans Large-Scale Multilingual Machine Translation System for WMT21 Nov 1, 2021 Knowledge Distillation Machine Translation
— Unverified 00 TernaryLLM: Ternarized Large Language Model Jun 11, 2024 Knowledge Distillation Language Modeling
— Unverified 00 Test-Time Adaptation Toward Personalized Speech Enhancement: Zero-Shot Learning with Knowledge Distillation May 8, 2021 Denoising Knowledge Distillation
— Unverified 00 Text is Text, No Matter What: Unifying Text Recognition using Knowledge Distillation Jul 26, 2021 Handwriting Recognition HTR
— Unverified 00 The Best of Both Worlds: Accurate Global and Personalized Models through Federated Learning with Data-Free Hyper-Knowledge Distillation Jan 21, 2023 Federated Learning Knowledge Distillation
— Unverified 00 The economic trade-offs of large language models: A case study Jun 8, 2023 Knowledge Distillation Prompt Engineering
— Unverified 00 The Estimation of Continual Causal Effect for Dataset Shifting Streams Apr 29, 2025 counterfactual Knowledge Distillation
— Unverified 00 The Graph's Apprentice: Teaching an LLM Low Level Knowledge for Circuit Quality Estimation Oct 30, 2024 Knowledge Distillation
— Unverified 00 The LMU Munich System for the WMT 2021 Large-Scale Multilingual Machine Translation Shared Task Nov 1, 2021 Data Augmentation Knowledge Distillation
— Unverified 00 The Microsoft Toolkit of Multi-Task Deep Neural Networks for Natural Language Understanding Feb 19, 2020 Knowledge Distillation Multi-Task Learning
— Unverified 00 The Mininglamp Machine Translation System for WMT21 Nov 1, 2021 Knowledge Distillation Machine Translation
— Unverified 00 The NiuTrans Machine Translation Systems for WMT19 Aug 1, 2019 Knowledge Distillation Machine Translation
— Unverified 00 The NiuTrans Machine Translation Systems for WMT21 Sep 22, 2021 Knowledge Distillation Machine Translation
— Unverified 00 The NiuTrans Machine Translation Systems for WMT20 Nov 1, 2020 Knowledge Distillation Machine Translation
— Unverified 00 The NiuTrans System for the WMT 2021 Efficiency Task Nov 1, 2021 GPU Knowledge Distillation
— Unverified 00 The NLP Cookbook: Modern Recipes for Transformer based Deep Learning Architectures Mar 23, 2021 Information Retrieval Knowledge Distillation
— Unverified 00 Theoretical Guarantees for LT-TTD: A Unified Transformer-based Architecture for Two-Level Ranking Systems May 7, 2025 Computational Efficiency Knowledge Distillation
— Unverified 00 The Privileged Students: On the Value of Initialization in Multilingual Knowledge Distillation Jun 24, 2024 Knowledge Distillation
— Unverified 00 The RoyalFlush System for the WMT 2022 Efficiency Task Dec 3, 2022 Decoder GPU
— Unverified 00 The Staged Knowledge Distillation in Video Classification: Harmonizing Student Progress by a Complementary Weakly Supervised Framework Jul 11, 2023 Knowledge Distillation Pseudo Label
— Unverified 00 The Unreasonable Effectiveness of Fully-Connected Layers for Low-Data Regimes Oct 11, 2022 Active Learning Knowledge Distillation
— Unverified 00 The USYD-JD Speech Translation System for IWSLT 2021 Jul 24, 2021 Knowledge Distillation NMT
— Unverified 00 The USYD-JD Speech Translation System for IWSLT2021 Aug 1, 2021 Knowledge Distillation NMT
— Unverified 00 The Xiaomi Text-to-Text Simultaneous Speech Translation System for IWSLT 2022 May 1, 2022 Automatic Speech Recognition Automatic Speech Recognition (ASR)
— Unverified 00 Three Factors to Improve Out-of-Distribution Detection Aug 2, 2023 Contrastive Learning Knowledge Distillation
— Unverified 00 TIMA: Text-Image Mutual Awareness for Balancing Zero-Shot Adversarial Robustness and Generalization Ability May 27, 2024 Adversarial Robustness Knowledge Distillation
— Unverified 00 TimeDistill: Efficient Long-Term Time Series Forecasting with MLP via Cross-Architecture Distillation Feb 20, 2025 Data Augmentation Knowledge Distillation
— Unverified 00 TinyM^2Net-V3: Memory-Aware Compressed Multimodal Deep Neural Networks for Sustainable Edge Deployment May 20, 2024 Knowledge Distillation Model Compression
— Unverified 00 XtremeDistil: Multi-stage Distillation for Massive Multilingual Models Apr 12, 2020 Knowledge Distillation named-entity-recognition
— Unverified 00 TinyViT: Fast Pretraining Distillation for Small Vision Transformers Jul 21, 2022 Image Classification Knowledge Distillation
— Unverified 00 TIP: Typifying the Interpretability of Procedures Jun 9, 2017 Knowledge Distillation
— Unverified 00 TKD: Temporal Knowledge Distillation for Active Perception Mar 4, 2019 Knowledge Distillation Object
— Unverified 00 ToDi: Token-wise Distillation via Fine-Grained Divergence Control May 22, 2025 Instruction Following Knowledge Distillation
— Unverified 00 TODM: Train Once Deploy Many Efficient Supernet-Based RNN-T Compression For On-device ASR Models Sep 5, 2023 Automatic Speech Recognition Automatic Speech Recognition (ASR)
— Unverified 00 Tokenizing Electron Cloud in Protein-Ligand Interaction Learning May 25, 2025 Knowledge Distillation Prediction
— Unverified 00 Token-Level Ensemble Distillation for Grapheme-to-Phoneme Conversion Apr 6, 2019 Automatic Speech Recognition Automatic Speech Recognition (ASR)
— Unverified 00 Topic Modeling for Maternal Health Using Reddit Apr 1, 2021 Knowledge Distillation
— Unverified 00 Topological Persistence Guided Knowledge Distillation for Wearable Sensor Data Jul 7, 2024 Activity Recognition Deep Learning
— Unverified 00 Topology Distillation for Recommender System Jun 16, 2021 Knowledge Distillation Model Compression
— Unverified 00 torchdistill: A Modular, Configuration-Driven Framework for Knowledge Distillation Nov 25, 2020 Image Classification Instance Segmentation
— Unverified 00 torchdistill Meets Hugging Face Libraries for Reproducible, Coding-Free Deep Learning Studies: A Case Study on NLP Oct 26, 2023 image-classification Image Classification
— Unverified 00 To Smooth or not to Smooth? On Compatibility between Label Smoothing and Knowledge Distillation Sep 29, 2021 image-classification Image Classification
— Unverified 00 Toward Data-centric Directed Graph Learning: An Entropy-driven Approach May 2, 2025 Graph Learning Knowledge Distillation
— Unverified 00