Compressing Convolutional Neural Networks via Factorized Convolutional Filters Jun 1, 2019 Model Compression
Code Code Available 05 Exploiting Kernel Sparsity and Entropy for Interpretable CNN Compression Dec 11, 2018 Clustering Model Compression
Code Code Available 05 FedSynth: Gradient Compression via Synthetic Data in Federated Learning Apr 4, 2022 Federated Learning Model Compression
Code Code Available 05 Trio-ViT: Post-Training Quantization and Acceleration for Softmax-Free Efficient Vision Transformer May 6, 2024 Efficient ViTs Model Compression
Code Code Available 05 From Dense to Sparse: Contrastive Pruning for Better Pre-trained Language Model Compression Dec 14, 2021 Contrastive Learning Language Modeling
Code Code Available 05 Compressed models are NOT miniature versions of large models Jul 18, 2024 Adversarial Attack Model Compression
— Unverified 00 Artemis: HE-Aware Training for Efficient Privacy-Preserving Machine Learning Oct 2, 2023 Model Compression Privacy Preserving
— Unverified 00 Comprehensive Survey of Model Compression and Speed up for Vision Transformers Apr 16, 2024 Computational Efficiency Edge-computing
— Unverified 00 Are We There Yet? A Measurement Study of Efficiency for LLM Applications on Mobile Devices Mar 10, 2025 CPU GPU
— Unverified 00 Comprehensive Study on Performance Evaluation and Optimization of Model Compression: Bridging Traditional Deep Learning and Large Language Models Jul 22, 2024 Deep Learning image-classification
— Unverified 00 ESPACE: Dimensionality Reduction of Activations for Model Compression Oct 7, 2024 Dimensionality Reduction model
— Unverified 00 Compositionality Unlocks Deep Interpretable Models Apr 3, 2025 Model Compression Tensor Networks
— Unverified 00 A Comprehensive Review and a Taxonomy of Edge Machine Learning: Requirements, Paradigms, and Techniques Feb 16, 2023 Edge-computing Model Compression
— Unverified 00 Accelerating Very Deep Convolutional Networks for Classification and Detection May 26, 2015 Classification General Classification
— Unverified 00 EPSD: Early Pruning with Self-Distillation for Efficient Model Compression Jan 31, 2024 Knowledge Distillation Model Compression
— Unverified 00 EoRA: Training-free Compensation for Compressed LLM with Eigenspace Low-Rank Approximation Oct 28, 2024 ARC Math
— Unverified 00 CompMarkGS: Robust Watermarking for Compressed 3D Gaussian Splatting Mar 17, 2025 3DGS 3D Reconstruction
— Unverified 00 Ensemble-Compression: A New Method for Parallel Training of Deep Neural Networks Jun 2, 2016 Model Compression
— Unverified 00 Enhancing Targeted Attack Transferability via Diversified Weight Pruning Aug 18, 2022 Diversity Model Compression
— Unverified 00 Complexity-Driven CNN Compression for Resource-constrained Edge AI Aug 26, 2022 Computational Efficiency Model Compression
— Unverified 00 Architecture Compression Feb 8, 2019 Decoder Model Compression
— Unverified 00 Compacting Deep Neural Networks for Internet of Things: Methods and Applications Mar 20, 2021 Diversity Knowledge Distillation
— Unverified 00 Enhanced Sparsification via Stimulative Training Mar 11, 2024 Knowledge Distillation Model Compression
— Unverified 00 CompactifAI: Extreme Compression of Large Language Models using Quantum-Inspired Tensor Networks Jan 25, 2024 Model Compression Quantization
— Unverified 00 Energy-efficient Knowledge Distillation for Spiking Neural Networks Jun 14, 2021 Knowledge Distillation Model Compression
— Unverified 00 EncCluster: Scalable Functional Encryption in Federated Learning through Weight Clustering and Probabilistic Filters Jun 13, 2024 Federated Learning Model Compression
— Unverified 00 Compact CNN Structure Learning by Knowledge Distillation Apr 19, 2021 Knowledge Distillation Model Compression
— Unverified 00 A Progressive Sub-Network Searching Framework for Dynamic Inference Sep 11, 2020 Model Compression
— Unverified 00 A Deep Cascade Network for Unaligned Face Attribute Classification Sep 12, 2017 Attribute Classification
— Unverified 00 Accelerating Machine Learning Primitives on Commodity Hardware Oct 8, 2023 CPU Model Compression
— Unverified 00 Enabling Deep Learning on Edge Devices through Filter Pruning and Knowledge Transfer Jan 22, 2022 image-classification Image Classification
— Unverified 00 Energy-Efficient Model Compression and Splitting for Collaborative Inference Over Time-Varying Channels Jun 2, 2021 Collaborative Inference image-classification
— Unverified 00 Enabling All In-Edge Deep Learning: A Literature Review Apr 7, 2022 All Deep Learning
— Unverified 00 Enable Deep Learning on Mobile Devices: Methods, Systems, and Applications Apr 25, 2022 AutoML Deep Learning
— Unverified 00 Communication-Efficient Federated Learning with Adaptive Compression under Dynamic Bandwidth May 6, 2024 Federated Learning Model Compression
— Unverified 00 Enhancing Inference Efficiency of Large Language Models: Investigating Optimization Strategies and Architectural Innovations Apr 2, 2024 Model Compression
— Unverified 00 Empowering Edge Intelligence: A Comprehensive Survey on On-Device AI Models Mar 8, 2025 Edge-computing Model Compression
— Unverified 00 ELRT: Efficient Low-Rank Training for Compact Convolutional Neural Networks Jan 18, 2024 Low-rank compression Model Compression
— Unverified 00 Communication-Efficient Distributed Online Learning with Kernels Nov 28, 2019 Model Compression
— Unverified 00 A Privacy-Preserving-Oriented DNN Pruning and Mobile Acceleration Framework Mar 13, 2020 Model Compression Privacy Preserving
— Unverified 00 E-LANG: Energy-Based Joint Inferencing of Super and Swift Language Models Mar 1, 2022 Decision Making Model Compression
— Unverified 00 EPIM: Efficient Processing-In-Memory Accelerators based on Epitome Nov 12, 2023 Model Compression Neural Architecture Search
— Unverified 00 Efficient Transformer Knowledge Distillation: A Performance Review Nov 22, 2023 Knowledge Distillation Model Compression
— Unverified 00 Error-aware Quantization through Noise Tempering Dec 11, 2022 Model Compression Quantization
— Unverified 00 Approximability and Generalisation Mar 15, 2022 Learning Theory Model Compression
— Unverified 00 Every Expert Matters: Towards Effective Knowledge Distillation for Mixture-of-Experts Language Models Feb 18, 2025 Knowledge Distillation Mixture-of-Experts
— Unverified 00 Efficient Supernet Training with Orthogonal Softmax for Scalable ASR Model Compression Jan 31, 2025 Model Compression
— Unverified 00 Efficient Speech Representation Learning with Low-Bit Quantization Dec 14, 2022 Model Compression Quantization
— Unverified 00 Efficient Recurrent Neural Networks using Structured Matrices in FPGAs Mar 20, 2018 Model Compression Time Series
— Unverified 00 CoLLD: Contrastive Layer-to-layer Distillation for Compressing Multilingual Pre-trained Speech Encoders Sep 14, 2023 Contrastive Learning Knowledge Distillation
— Unverified 00