GraphKD: Exploring Knowledge Distillation Towards Document Object Detection with Structured Graph Creation Feb 17, 2024 Knowledge Distillation object-detection
Code Code Available 1On Good Practices for Task-Specific Distillation of Large Pretrained Visual Models Feb 17, 2024 Data Augmentation Knowledge Distillation
— Unverified 0Knowledge Distillation Based on Transformed Teacher Matching Feb 17, 2024 Knowledge Distillation
Code Code Available 1BitDistiller: Unleashing the Potential of Sub-4-Bit LLMs via Self-Distillation Feb 16, 2024 Knowledge Distillation Quantization
Code Code Available 4Incremental Sequence Labeling: A Tale of Two Shifts Feb 16, 2024 Incremental Learning Knowledge Distillation
Code Code Available 2Cultural Commonsense Knowledge for Intercultural Dialogues Feb 16, 2024 Knowledge Distillation Specificity
— Unverified 0FedD2S: Personalized Data-Free Federated Knowledge Distillation Feb 16, 2024 Data-free Knowledge Distillation Fairness
— Unverified 0Distilled Gradual Pruning with Pruned Fine-tuning Feb 15, 2024 Image Classification Knowledge Distillation
Code Code Available 0Walsh-domain Neural Network for Power Amplifier Behavioral Modelling and Digital Predistortion Feb 15, 2024 Knowledge Distillation
— Unverified 0Model Compression and Efficient Inference for Large Language Models: A Survey Feb 15, 2024 Knowledge Distillation Model Compression
— Unverified 0NutePrune: Efficient Progressive Pruning with Numerous Teachers for Large Language Models Feb 15, 2024 Knowledge Distillation
Code Code Available 0Leveraging Large Language Models for Enhanced NLP Task Performance through Knowledge Distillation and Optimized Training Strategies Feb 14, 2024 Knowledge Distillation named-entity-recognition
— Unverified 0FedSiKD: Clients Similarity and Knowledge Distillation: Addressing Non-i.i.d. and Constraints in Federated Learning Feb 14, 2024 Federated Learning Knowledge Distillation
Code Code Available 0Integrating ChatGPT into Secure Hospital Networks: A Case Study on Improving Radiology Report Analysis Feb 14, 2024 Contrastive Learning Knowledge Distillation
— Unverified 0APALU: A Trainable, Adaptive Activation Function for Deep Learning Networks Feb 13, 2024 Anomaly Detection Deep Learning
— Unverified 0Two-Stage Multi-task Self-Supervised Learning for Medical Image Segmentation Feb 11, 2024 Auxiliary Learning Image Segmentation
— Unverified 0Training Heterogeneous Client Models using Knowledge Distillation in Serverless Federated Learning Feb 11, 2024 Federated Learning Knowledge Distillation
Code Code Available 1Domain Adaptable Fine-Tune Distillation Framework For Advancing Farm Surveillance Feb 10, 2024 Computational Efficiency Knowledge Distillation
Code Code Available 0Embedding Compression for Teacher-to-Student Knowledge Transfer Feb 9, 2024 Knowledge Distillation Transfer Learning
— Unverified 0Multi-source-free Domain Adaptation via Uncertainty-aware Adaptive Distillation Feb 9, 2024 Domain Adaptation Knowledge Distillation
Code Code Available 0Large Language Model Meets Graph Neural Network in Knowledge Distillation Feb 8, 2024 Contrastive Learning Graph Attention
— Unverified 0Knowledge Distillation for Road Detection based on cross-model Semi-Supervised Learning Feb 7, 2024 Knowledge Distillation Road Segmentation
— Unverified 0Beyond Answers: Transferring Reasoning Capabilities to Smaller LLMs Using Multi-Teacher Knowledge Distillation Feb 7, 2024 Diversity Knowledge Distillation
Code Code Available 0EfficientViT-SAM: Accelerated Segment Anything Model Without Accuracy Loss Feb 7, 2024 Decoder GPU
— Unverified 0Vision Superalignment: Weak-to-Strong Generalization for Vision Foundation Models Feb 6, 2024 Few-Shot Learning Knowledge Distillation
Code Code Available 1DistiLLM: Towards Streamlined Distillation for Large Language Models Feb 6, 2024 Instruction Following Knowledge Distillation
Code Code Available 3A Survey on Transformer Compression Feb 5, 2024 Knowledge Distillation Mamba
— Unverified 0Large Language Model Distilling Medication Recommendation Model Feb 5, 2024 Knowledge Distillation Language Modeling
Code Code Available 1BGE M3-Embedding: Multi-Lingual, Multi-Functionality, Multi-Granularity Text Embeddings Through Self-Knowledge Distillation Feb 5, 2024 Knowledge Distillation Retrieval
Code Code Available 1Dual Knowledge Distillation for Efficient Sound Event Detection Feb 5, 2024 Event Detection Knowledge Distillation
— Unverified 0Good Teachers Explain: Explanation-Enhanced Knowledge Distillation Feb 5, 2024 Knowledge Distillation
Code Code Available 1LQER: Low-Rank Quantization Error Reconstruction for LLMs Feb 4, 2024 Knowledge Distillation Quantization
Code Code Available 1Cooperative Knowledge Distillation: A Learner Agnostic Approach Feb 2, 2024 counterfactual Knowledge Distillation
Code Code Available 0Bi-CryptoNets: Leveraging Different-Level Privacy for Encrypted Inference Feb 2, 2024 Knowledge Distillation Privacy Preserving
— Unverified 0Spiking CenterNet: A Distillation-boosted Spiking Neural Network for Object Detection Feb 2, 2024 Decoder Knowledge Distillation
— Unverified 0Class incremental learning with probability dampening and cascaded gated classifier Feb 2, 2024 class-incremental learning Class Incremental Learning
Code Code Available 0Faster Inference of Integer SWIN Transformer by Removing the GELU Activation Feb 2, 2024 GPU image-classification
— Unverified 0Addressing Bias Through Ensemble Learning and Regularized Fine-Tuning Feb 1, 2024 Ensemble Learning Knowledge Distillation
— Unverified 0Dual-Student Knowledge Distillation Networks for Unsupervised Anomaly Detection Feb 1, 2024 Anomaly Detection Anomaly Segmentation
— Unverified 0Augmenting Offline Reinforcement Learning with State-only Interactions Feb 1, 2024 D4RL Data Augmentation
— Unverified 0Scavenging Hyena: Distilling Transformers into Long Convolution Models Jan 31, 2024 Knowledge Distillation
— Unverified 0EPSD: Early Pruning with Self-Distillation for Efficient Model Compression Jan 31, 2024 Knowledge Distillation Model Compression
— Unverified 0LLaMP: Large Language Model Made Powerful for High-fidelity Materials Knowledge Retrieval and Distillation Jan 30, 2024 Hallucination Knowledge Distillation
Code Code Available 2Stolen Subwords: Importance of Vocabularies for Machine Translation Model Stealing Jan 29, 2024 Knowledge Distillation Machine Translation
Code Code Available 0TQCompressor: improving tensor decomposition methods in neural networks via permutations Jan 29, 2024 Knowledge Distillation Model Compression
Code Code Available 0Face to Cartoon Incremental Super-Resolution using Knowledge Distillation Jan 27, 2024 Hallucination Incremental Learning
— Unverified 0Dynamic Transformer Architecture for Continual Learning of Multimodal Tasks Jan 27, 2024 Continual Learning Edge-computing
— Unverified 0Distilling Privileged Multimodal Information for Expression Recognition using Optimal Transport Jan 27, 2024 Diversity Knowledge Distillation
— Unverified 0A Comprehensive Survey of Compression Algorithms for Language Models Jan 27, 2024 Knowledge Distillation Quantization
— Unverified 0Large Language Model Guided Knowledge Distillation for Time Series Anomaly Detection Jan 26, 2024 Anomaly Detection Knowledge Distillation
— Unverified 0