Is Modularity Transferable? A Case Study through the Lens of Knowledge Distillation Mar 27, 2024 Domain Adaptation Knowledge Distillation
Code Code Available 0Dense Vision Transformer Compression with Few Samples Mar 27, 2024 Model Compression
— Unverified 0Are Compressed Language Models Less Subgroup Robust? Mar 26, 2024 Model Compression
Code Code Available 0Order of Compression: A Systematic and Optimal Sequence to Combinationally Compress CNN Mar 26, 2024 Knowledge Distillation Model Compression
— Unverified 0Tiny Models are the Computational Saver for Large Models Mar 26, 2024 Computational Efficiency Image Classification
Code Code Available 0Magic for the Age of Quantized DNNs Mar 22, 2024 Model Compression Quantization
— Unverified 0Advancing IIoT with Over-the-Air Federated Learning: The Role of Iterative Magnitude Pruning Mar 21, 2024 Federated Learning Model Compression
— Unverified 0DiPaCo: Distributed Path Composition Mar 15, 2024 Language Modelling Model Compression
— Unverified 0BRIEDGE: EEG-Adaptive Edge AI for Multi-Brain to Multi-Robot Interaction Mar 14, 2024 EEG Model Compression
— Unverified 0Adversarial Fine-tuning of Compressed Neural Networks for Joint Improvement of Robustness and Efficiency Mar 14, 2024 Adversarial Robustness Model Compression
Code Code Available 0Maxwell's Demon at Work: Efficient Pruning by Leveraging Saturation of Neurons Mar 12, 2024 Continual Learning Model Compression
— Unverified 0Enhanced Sparsification via Stimulative Training Mar 11, 2024 Knowledge Distillation Model Compression
— Unverified 0Optimal Policy Sparsification and Low Rank Decomposition for Deep Reinforcement Learning Mar 10, 2024 Deep Reinforcement Learning Edge-computing
— Unverified 0DyCE: Dynamically Configurable Exiting for Deep Learning Compression and Real-time Scaling Mar 4, 2024 image-classification Image Classification
Code Code Available 0Towards efficient deep autoencoders for multivariate time series anomaly detection Mar 4, 2024 Anomaly Detection Model Compression
— Unverified 0Differentially Private Knowledge Distillation via Synthetic Text Generation Mar 1, 2024 Knowledge Distillation Model Compression
Code Code Available 0Model Compression Method for S4 with Diagonal State Space Layers using Balanced Truncation Feb 25, 2024 Model Compression
— Unverified 0FinGPT-HPC: Efficient Pretraining and Finetuning Large Language Models for Financial Applications with High-Performance Computing Feb 21, 2024 GPU Model Compression
— Unverified 0From Cloud to Edge: Rethinking Generative AI for Low-Resource Design Challenges Feb 20, 2024 Edge-computing Model Compression
— Unverified 0Towards a tailored mixed-precision sub-8-bit quantization scheme for Gated Recurrent Units using Genetic Algorithms Feb 19, 2024 Model Compression Quantization
— Unverified 0Extraction of nonlinearity in neural networks with Koopman operator Feb 18, 2024 Model Compression
— Unverified 0Model Compression and Efficient Inference for Large Language Models: A Survey Feb 15, 2024 Knowledge Distillation Model Compression
— Unverified 0Bayesian Deep Learning Via Expectation Maximization and Turbo Deep Approximate Message Passing Feb 12, 2024 Bayesian Inference Federated Learning
— Unverified 0Memory-Efficient Vision Transformers: An Activation-Aware Mixed-Rank Compression Strategy Feb 8, 2024 Model Compression
— Unverified 0L4Q: Parameter Efficient Quantization-Aware Fine-Tuning on Large Language Models Feb 7, 2024 Few-Shot Learning In-Context Learning
— Unverified 0Expediting In-Network Federated Learning by Voting-Based Consensus Model Compression Feb 6, 2024 Federated Learning Model Compression
— Unverified 0The Potential of AutoML for Recommender Systems Feb 6, 2024 AutoML Machine Translation
— Unverified 0Fed-CVLC: Compressing Federated Learning Communications with Variable-Length Codes Feb 6, 2024 Federated Learning Model Compression
— Unverified 0Dynamic Sparse Learning: A Novel Paradigm for Efficient Recommendation Feb 5, 2024 Model Compression Recommendation Systems
— Unverified 0A Survey on Transformer Compression Feb 5, 2024 Knowledge Distillation Mamba
— Unverified 0Mobile Fitting Room: On-device Virtual Try-on via Diffusion Models Feb 2, 2024 Image Generation Model Compression
— Unverified 0EPSD: Early Pruning with Self-Distillation for Efficient Model Compression Jan 31, 2024 Knowledge Distillation Model Compression
— Unverified 0Effective Multi-Stage Training Model For Edge Computing Devices In Intrusion Detection Jan 31, 2024 Edge-computing Intrusion Detection
— Unverified 0RADIN: Souping on a Budget Jan 31, 2024 Ensemble Learning Model Compression
— Unverified 0Diffusion Model Compression for Image-to-Image Translation Jan 31, 2024 Conditional Image Generation Denoising
— Unverified 0SwapNet: Efficient Swapping for DNN Inference on Edge AI Devices Beyond the Memory Budget Jan 30, 2024 GPU Model Compression
— Unverified 0TQCompressor: improving tensor decomposition methods in neural networks via permutations Jan 29, 2024 Knowledge Distillation Model Compression
Code Code Available 0CompactifAI: Extreme Compression of Large Language Models using Quantum-Inspired Tensor Networks Jan 25, 2024 Model Compression Quantization
— Unverified 0Large receptive field strategy and important feature extraction strategy in 3D object detection Jan 22, 2024 3D Object Detection Autonomous Driving
— Unverified 0Pruning for Protection: Increasing Jailbreak Resistance in Aligned LLMs Without Fine-Tuning Jan 19, 2024 Model Compression
Code Code Available 0ELRT: Efficient Low-Rank Training for Compact Convolutional Neural Networks Jan 18, 2024 Low-rank compression Model Compression
— Unverified 0Model Compression Techniques in Biometrics Applications: A Survey Jan 18, 2024 Fairness Knowledge Distillation
Code Code Available 0SymbolNet: Neural Symbolic Regression with Adaptive Dynamic Pruning for Compression Jan 18, 2024 Jet Tagging Model Compression
Code Code Available 0Convolutional Neural Network Compression via Dynamic Parameter Rank Pruning Jan 15, 2024 Model Compression Neural Network Compression
— Unverified 0Knowledge Translation: A New Pathway for Model Compression Jan 11, 2024 Data Augmentation model
Code Code Available 0FFSplit: Split Feed-Forward Network For Optimizing Accuracy-Efficiency Trade-off in Language Model Inference Jan 8, 2024 GPU Language Modeling
— Unverified 0Understanding LLMs: A Comprehensive Overview from Training to Inference Jan 4, 2024 Language Modeling Language Modelling
— Unverified 0Safety and Performance, Why Not Both? Bi-Objective Optimized Model Compression against Heterogeneous Attacks Toward AI Software Deployment Jan 2, 2024 Inference Attack Membership Inference Attack
Code Code Available 0Unleashing Channel Potential: Space-Frequency Selection Convolution for SAR Object Detection Jan 1, 2024 feature selection Model Compression
— Unverified 0Data-Free Quantization via Pseudo-label Filtering Jan 1, 2024 Data Free Quantization Model Compression
— Unverified 0