Distilled Split Deep Neural Networks for Edge-Assisted Real-Time Systems Oct 1, 2019 Edge-computing Image Classification
Code Code Available 1Examining Post-Training Quantization for Mixture-of-Experts: A Benchmark Jun 12, 2024 Benchmarking Mixture-of-Experts
Code Code Available 1Faster and Lighter LLMs: A Survey on Current Challenges and Way Forward Feb 2, 2024 Model Compression Survey
Code Code Available 1Fast Vocabulary Transfer for Language Model Compression Feb 15, 2024 Language Modeling Language Modelling
Code Code Available 1A Real-time Low-cost Artificial Intelligence System for Autonomous Spraying in Palm Plantations Mar 6, 2021 Model Compression Navigate
Code Code Available 1FIMA-Q: Post-Training Quantization for Vision Transformers by Fisher Information Matrix Approximation Jun 13, 2025 Model Compression Quantization
Code Code Available 1Forget the Data and Fine-Tuning! Just Fold the Network to Compress Feb 14, 2025 Model Compression
Code Code Available 1Gaussian RAM: Lightweight Image Classification via Stochastic Retina-Inspired Glimpse and Reinforcement Learning Nov 12, 2020 Classification General Classification
Code Code Available 1DS-Net++: Dynamic Weight Slicing for Efficient Inference in CNNs and Transformers Sep 21, 2021 Fairness Model Compression
Code Code Available 1AD-KD: Attribution-Driven Knowledge Distillation for Language Model Compression May 17, 2023 Knowledge Distillation Language Modeling
Code Code Available 1ADMM-NN: An Algorithm-Hardware Co-Design Framework of DNNs Using Alternating Direction Method of Multipliers Dec 31, 2018 Model Compression Quantization
Code Code Available 1Global Sparse Momentum SGD for Pruning Very Deep Neural Networks Sep 27, 2019 Model Compression
Code Code Available 1Discovering Dynamic Patterns from Spatiotemporal Data with Time-Varying Low-Rank Autoregression Nov 28, 2022 Model Compression
Code Code Available 1DE-RRD: A Knowledge Distillation Framework for Recommender System Dec 8, 2020 Knowledge Distillation Model Compression
Code Code Available 1Discrimination-aware Channel Pruning for Deep Neural Networks Oct 28, 2018 channel selection Model Compression
Code Code Available 1Dual Relation Knowledge Distillation for Object Detection Feb 11, 2023 Knowledge Distillation Model Compression
Code Code Available 1DarwinLM: Evolutionary Structured Pruning of Large Language Models Feb 11, 2025 Model Compression
Code Code Available 1Contrastive Representation Distillation Oct 23, 2019 Contrastive Learning Knowledge Distillation
Code Code Available 1CPrune: Compiler-Informed Model Pruning for Efficient Target-Aware DNN Execution Jul 4, 2022 Compiler Optimization image-classification
Code Code Available 1Designing Large Foundation Models for Efficient Training and Inference: A Survey Sep 3, 2024 Knowledge Distillation Model Compression
Code Code Available 1Computation-Efficient Knowledge Distillation via Uncertainty-Aware Mixup Dec 17, 2020 Informativeness Knowledge Distillation
Code Code Available 1Contrastive Distillation on Intermediate Representations for Language Model Compression Sep 29, 2020 Knowledge Distillation Language Modeling
Code Code Available 1Data-Free Network Quantization With Adversarial Knowledge Distillation May 8, 2020 Knowledge Distillation Model Compression
Code Code Available 1Dynamic Channel Pruning: Feature Boosting and Suppression Oct 12, 2018 Model Compression Network Pruning
Code Code Available 1Comprehensive Knowledge Distillation with Causal Intervention Dec 1, 2021 Causal Inference Knowledge Distillation
Code Code Available 1Aligned Structured Sparsity Learning for Efficient Image Super-Resolution Dec 1, 2021 Image Super-Resolution Knowledge Distillation
Code Code Available 1Streamlining Redundant Layers to Compress Large Language Models Mar 28, 2024 Model Compression
Code Code Available 1Compacting, Picking and Growing for Unforgetting Continual Learning Oct 15, 2019 Age And Gender Classification Continual Learning
Code Code Available 1Composable Interventions for Language Models Jul 9, 2024 knowledge editing Machine Unlearning
Code Code Available 1An Information Theory-inspired Strategy for Automatic Network Pruning Aug 19, 2021 AutoML Model Compression
Code Code Available 1Consistent Quantity-Quality Control across Scenes for Deployment-Aware Gaussian Splatting May 15, 2025 3DGS Model Compression
Code Code Available 1Constraint-aware and Ranking-distilled Token Pruning for Efficient Transformer Inference Jun 26, 2023 CPU Model Compression
Code Code Available 1Communication-Efficient Diffusion Strategy for Performance Improvement of Federated Learning with Non-IID Data Jul 15, 2022 Federated Learning Model Compression
Code Code Available 1Communication-Computation Trade-Off in Resource-Constrained Edge Inference Jun 3, 2020 Edge-computing Model Compression
Code Code Available 1Activation-Informed Merging of Large Language Models Feb 4, 2025 Computational Efficiency Continual Learning
Code Code Available 1CrossKD: Cross-Head Knowledge Distillation for Object Detection Jun 20, 2023 Dense Object Detection Knowledge Distillation
Code Code Available 1Deep Compression for PyTorch Model Deployment on Microcontrollers Mar 29, 2021 model Model Compression
Code Code Available 1Densely Guided Knowledge Distillation using Multiple Teacher Assistants Sep 18, 2020 Knowledge Distillation Model Compression
Code Code Available 1Differentiable Model Compression via Pseudo Quantization Noise Apr 20, 2021 Audio Source Separation image-classification
Code Code Available 1Communication-Efficient Federated Learning through Adaptive Weight Clustering and Server-Side Distillation Jan 25, 2024 Clustering Federated Learning
Code Code Available 1Discrimination-aware Network Pruning for Deep Model Compression Jan 4, 2020 Face Recognition image-classification
Code Code Available 1DiSparse: Disentangled Sparsification for Multitask Model Compression Jun 9, 2022 model Model Compression
Code Code Available 1Distilling Linguistic Context for Language Model Compression Sep 17, 2021 Knowledge Distillation Language Modeling
Code Code Available 1Distilling Object Detectors with Feature Richness Nov 1, 2021 Knowledge Distillation Model Compression
Code Code Available 1CoA: Towards Real Image Dehazing via Compression-and-Adaptation Jan 1, 2025 Image Dehazing Model Compression
Code Code Available 1EarlyBERT: Efficient BERT Training via Early-bird Lottery Tickets Dec 31, 2020 Model Compression
Code Code Available 1An Efficient Multilingual Language Model Compression through Vocabulary Trimming May 24, 2023 Language Modeling Language Modelling
Code Code Available 1Efficient On-Device Session-Based Recommendation Sep 27, 2022 Knowledge Distillation Model Compression
Code Code Available 1Learning Efficient Vision Transformers via Fine-Grained Manifold Distillation Jul 3, 2021 Knowledge Distillation Model Compression
Code Code Available 1Model LEGO: Creating Models Like Disassembling and Assembling Building Blocks Mar 25, 2022 Incremental Learning Knowledge Distillation
Code Code Available 1