USM-Lite: Quantization and Sparsity Aware Fine-tuning for Speech Recognition with Universal Speech Models Dec 13, 2023 Automatic Speech Recognition Automatic Speech Recognition (ASR)
— Unverified 0Rethinking Compression: Reduced Order Modelling of Latent Features in Large Language Models Dec 12, 2023 GPU Model Compression
Code Code Available 1Neural Architecture Codesign for Fast Bragg Peak Analysis Dec 10, 2023 AutoML Model Compression
— Unverified 0Large Multimodal Model Compression via Efficient Pruning and Distillation at AntGroup Dec 10, 2023 Model Compression
Code Code Available 0Understanding the Effect of Model Compression on Social Bias in Large Language Models Dec 9, 2023 Knowledge Distillation Model Compression
Code Code Available 0Language Model Knowledge Distillation for Efficient Question Answering in Spanish Dec 7, 2023 Knowledge Distillation Language Modeling
Code Code Available 0Physics Inspired Criterion for Pruning-Quantization Joint Learning Dec 1, 2023 image-classification Image Classification
Code Code Available 0The Efficiency Spectrum of Large Language Models: An Algorithmic Survey Dec 1, 2023 Model Compression Survey
Code Code Available 0LayerCollapse: Adaptive compression of neural networks Nov 29, 2023 Computational Efficiency image-classification
— Unverified 0Privacy and Accuracy Implications of Model Complexity and Integration in Heterogeneous Federated Learning Nov 29, 2023 Federated Learning Inference Attack
Code Code Available 0Towards Higher Ranks via Adversarial Weight Pruning Nov 29, 2023 Model Compression Network Pruning
Code Code Available 0Relationship between Model Compression and Adversarial Robustness: A Review of Current Evidence Nov 27, 2023 Adversarial Robustness Model Compression
— Unverified 0Cosine Similarity Knowledge Distillation for Individual Class Information Transfer Nov 24, 2023 Knowledge Distillation Model Compression
— Unverified 0Knowledge Distillation Based Semantic Communications For Multiple Users Nov 23, 2023 Decoder Knowledge Distillation
— Unverified 0Education distillation:getting student models to learn in shcools Nov 23, 2023 Incremental Learning Knowledge Distillation
— Unverified 0Efficient Transformer Knowledge Distillation: A Performance Review Nov 22, 2023 Knowledge Distillation Model Compression
— Unverified 0Compact 3D Gaussian Representation for Radiance Field Nov 22, 2023 3DGS Model Compression
Code Code Available 2Towards Better Parameter-Efficient Fine-Tuning for Large Language Models: A Position Paper Nov 22, 2023 Model Compression parameter-efficient fine-tuning
— Unverified 0Shedding the Bits: Pushing the Boundaries of Quantization with Minifloats on FPGAs Nov 21, 2023 Model Compression Quantization
— Unverified 0LQ-LoRA: Low-rank Plus Quantized Matrix Decomposition for Efficient Language Model Finetuning Nov 20, 2023 GPU Language Modeling
Code Code Available 1Efficient Neural Networks for Tiny Machine Learning: A Comprehensive Review Nov 20, 2023 Model Compression Quantization
— Unverified 0On the Impact of Calibration Data in Post-training Quantization and Pruning Nov 16, 2023 Model Compression Quantization
— Unverified 0A Speed Odyssey for Deployable Quantization of LLMs Nov 16, 2023 Language Modeling Language Modelling
— Unverified 0FedCode: Communication-Efficient Federated Learning via Transferring Codebooks Nov 15, 2023 Federated Learning Model Compression
— Unverified 0EPIM: Efficient Processing-In-Memory Accelerators based on Epitome Nov 12, 2023 Model Compression Neural Architecture Search
— Unverified 0What is Lost in Knowledge Distillation? Nov 7, 2023 Knowledge Distillation Model Compression
— Unverified 0Supervised domain adaptation for building extraction from off-nadir aerial images Nov 7, 2023 Domain Adaptation Earth Observation
— Unverified 0Asymmetric Masked Distillation for Pre-Training Small Foundation Models Nov 6, 2023 Action Classification Action Recognition
Code Code Available 0Data-Free Distillation of Language Model by Text-to-Text Transfer Nov 3, 2023 Data-free Knowledge Distillation Diversity
— Unverified 0Divergent Token Metrics: Measuring degradation to prune away LLM components -- and optimize quantization Nov 2, 2023 Management Model Compression
— Unverified 0Retrieval-based Knowledge Transfer: An Effective Approach for Extreme Large Language Model Compression Oct 24, 2023 Language Modeling Language Modelling
— Unverified 0LXMERT Model Compression for Visual Question Answering Oct 23, 2023 model Model Compression
Code Code Available 0Data-Free Knowledge Distillation Using Adversarially Perturbed OpenGL Shader Images Oct 20, 2023 Data Augmentation Data-free Knowledge Distillation
— Unverified 0In defense of parameter sharing for model-compression Oct 17, 2023 Model Compression
— Unverified 0USDC: Unified Static and Dynamic Compression for Visual Transformer Oct 17, 2023 Model Compression
— Unverified 0Efficient Apple Maturity and Damage Assessment: A Lightweight Detection Model with GAN and Attention Mechanism Oct 13, 2023 Decision Making Model Compression
— Unverified 0What do larger image classifiers memorise? Oct 9, 2023 image-classification Image Classification
— Unverified 0Accelerating Machine Learning Primitives on Commodity Hardware Oct 8, 2023 CPU Model Compression
— Unverified 0A Corrected Expected Improvement Acquisition Function Under Noisy Observations Oct 8, 2023 Bayesian Optimization Model Compression
Code Code Available 0Model Compression in Practice: Lessons Learned from Practitioners Creating On-device Machine Learning Experiences Oct 6, 2023 Model Compression
— Unverified 0Robustness-Guided Image Synthesis for Data-Free Quantization Oct 5, 2023 Data Free Quantization Diversity
— Unverified 0Sparse Deep Learning for Time Series Data: Theory and Applications Oct 5, 2023 Deep Learning Model Compression
— Unverified 0ECoFLaP: Efficient Coarse-to-Fine Layer-Wise Pruning for Vision-Language Models Oct 4, 2023 Model Compression
— Unverified 0Sweeping Heterogeneity with Smart MoPs: Mixture of Prompts for LLM Task Adaptation Oct 4, 2023 Model Compression Text Summarization
— Unverified 0Artemis: HE-Aware Training for Efficient Privacy-Preserving Machine Learning Oct 2, 2023 Model Compression Privacy Preserving
— Unverified 0Bridging the Gap Between Foundation Models and Heterogeneous Federated Learning Sep 30, 2023 Federated Learning Model Compression
— Unverified 0Distilling Inductive Bias: Knowledge Distillation Beyond Model Compression Sep 30, 2023 Inductive Bias Knowledge Distillation
— Unverified 0CAIT: Triple-Win Compression towards High Accuracy, Fast Inference, and Favorable Transferability For ViTs Sep 27, 2023 Model Compression Semantic Segmentation
— Unverified 0On the Impact of Quantization and Pruning of Self-Supervised Speech Models for Downstream Speech Recognition Tasks "In-the-Wild'' Sep 25, 2023 Data Augmentation Model Compression
— Unverified 0VIC-KD: Variance-Invariance-Covariance Knowledge Distillation to Make Keyword Spotting More Robust Against Adversarial Attacks Sep 22, 2023 Adversarial Robustness Keyword Spotting
— Unverified 0