On the Impact of Quantization and Pruning of Self-Supervised Speech Models for Downstream Speech Recognition Tasks "In-the-Wild'' Sep 25, 2023 Data Augmentation Model Compression
— Unverified 0VIC-KD: Variance-Invariance-Covariance Knowledge Distillation to Make Keyword Spotting More Robust Against Adversarial Attacks Sep 22, 2023 Adversarial Robustness Keyword Spotting
— Unverified 0Pruning Large Language Models via Accuracy Predictor Sep 18, 2023 MMLU Model Compression
— Unverified 0Training dynamic models using early exits for automatic speech recognition on resource-constrained devices Sep 18, 2023 Automatic Speech Recognition Automatic Speech Recognition (ASR)
Code Code Available 0Two-Step Knowledge Distillation for Tiny Speech Enhancement Sep 15, 2023 Knowledge Distillation Model Compression
— Unverified 0CoLLD: Contrastive Layer-to-layer Distillation for Compressing Multilingual Pre-trained Speech Encoders Sep 14, 2023 Contrastive Learning Knowledge Distillation
— Unverified 0Training Acceleration of Low-Rank Decomposed Networks using Sequential Freezing and Rank Quantization Sep 7, 2023 Model Compression Quantization
— Unverified 0Norm Tweaking: High-performance Low-bit Quantization of Large Language Models Sep 6, 2023 Model Compression Quantization
— Unverified 0Compressing Vision Transformers for Low-Resource Visual Learning Sep 5, 2023 Autonomous Navigation image-classification
Code Code Available 0ADC/DAC-Free Analog Acceleration of Deep Neural Networks with Frequency Transformation Sep 4, 2023 Computational Efficiency Model Compression
— Unverified 0Uncovering the Hidden Cost of Model Compression Aug 29, 2023 model Model Compression
Code Code Available 0Computation-efficient Deep Learning for Computer Vision: A Survey Aug 27, 2023 Autonomous Vehicles Deep Learning
— Unverified 0Improving Knowledge Distillation for BERT Models: Loss Functions, Mapping Methods, and Weight Tuning Aug 26, 2023 Knowledge Distillation Model Compression
— Unverified 0DLIP: Distilling Language-Image Pre-training Aug 24, 2023 Image Captioning Image-text Retrieval
— Unverified 0QD-BEV : Quantization-aware View-guided Distillation for Multi-view 3D Object Detection Aug 21, 2023 3D Object Detection Model Compression
— Unverified 0Learning Disentangled Representation with Mutual Information Maximization for Real-Time UAV Tracking Aug 20, 2023 CPU Model Compression
— Unverified 0SHARK: A Lightweight Model Compression Approach for Large-scale Recommender Systems Aug 18, 2023 Model Compression Quantization
— Unverified 0Spike-and-slab shrinkage priors for structurally sparse Bayesian neural networks Aug 17, 2023 Computational Efficiency Model Compression
— Unverified 0Benchmarking Adversarial Robustness of Compressed Deep Learning Models Aug 16, 2023 Adversarial Robustness Benchmarking
— Unverified 0Shortcut-V2V: Compression Framework for Video-to-Video Translation based on Temporal Redundancy Reduction Aug 15, 2023 Computational Efficiency Model Compression
— Unverified 0A Survey on Model Compression for Large Language Models Aug 15, 2023 Benchmarking Knowledge Distillation
— Unverified 0FedEdge AI-TC: A Semi-supervised Traffic Classification Method based on Trusted Federated Deep Learning for Mobile Edge Computing Aug 14, 2023 Edge-computing Federated Learning
— Unverified 0Resource Constrained Model Compression via Minimax Optimization for Spiking Neural Networks Aug 9, 2023 Model Compression Sparse Learning
Code Code Available 0Accurate Neural Network Pruning Requires Rethinking Sparse Optimization Aug 3, 2023 Model Compression Network Pruning
— Unverified 0MIMONet: Multi-Input Multi-Output On-Device Deep Learning Jul 22, 2023 Deep Learning Model Compression
— Unverified 0Model Compression Methods for YOLOv5: A Review Jul 21, 2023 Knowledge Distillation model
— Unverified 0Impact of Disentanglement on Pruning Neural Networks Jul 19, 2023 Disentanglement Model Compression
— Unverified 0Knowledge Distillation for Object Detection: from generic to remote sensing datasets Jul 18, 2023 Knowledge Distillation Model Compression
— Unverified 0CA-LoRA: Adapting Existing LoRA for Compressed LLMs to Enable Efficient Multi-Tasking on Personal Devices Jul 15, 2023 Model Compression
Code Code Available 0Distilling Universal and Joint Knowledge for Cross-Domain Model Compression on Time Series Data Jul 7, 2023 Knowledge Distillation Model Compression
Code Code Available 0Distilled Pruning: Using Synthetic Data to Win the Lottery Jul 7, 2023 Efficient Neural Network Model Compression
Code Code Available 0TensorGPT: Efficient Compression of Large Language Models based on Tensor-Train Decomposition Jul 2, 2023 Model Compression
— Unverified 0Data-Free Quantization via Mixed-Precision Compensation without Fine-Tuning Jul 2, 2023 Data Free Quantization Model Compression
— Unverified 0An Efficient Sparse Inference Software Accelerator for Transformer-based Language Models on CPUs Jun 28, 2023 Model Compression
— Unverified 0Low-Rank Prune-And-Factorize for Language Model Compression Jun 25, 2023 Language Modeling Language Modelling
— Unverified 0Feature Adversarial Distillation for Point Cloud Classification Jun 25, 2023 Classification FAD
— Unverified 0Partitioning-Guided K-Means: Extreme Empty Cluster Resolution for Extreme Model Compression Jun 24, 2023 Model Compression Quantization
— Unverified 0Data-Free Backbone Fine-Tuning for Pruned Neural Networks Jun 22, 2023 2D Human Pose Estimation image-classification
Code Code Available 0DynaQuant: Compressing Deep Learning Training Checkpoints via Dynamic Quantization Jun 20, 2023 Deep Learning Model Compression
— Unverified 0LoSparse: Structured Compression of Large Language Models based on Low-Rank and Sparse Approximation Jun 20, 2023 Diversity Model Compression
— Unverified 0Neural Network Compression using Binarization and Few Full-Precision Weights Jun 15, 2023 Binarization CPU
— Unverified 0Deep Model Compression Also Helps Models Capture Ambiguity Jun 12, 2023 model Model Compression
Code Code Available 0A Brief Review of Hypernetworks in Deep Learning Jun 12, 2023 Causal Inference Continual Learning
Code Code Available 0Riemannian Low-Rank Model Compression for Federated Learning with Over-the-Air Aggregation Jun 4, 2023 Federated Learning Model Compression
— Unverified 0Modular Transformers: Compressing Transformers into Modularized Layers for Flexible Efficient Inference Jun 4, 2023 Decoder Knowledge Distillation
— Unverified 0Low-Complexity Acoustic Scene Classification Using Data Augmentation and Lightweight ResNet Jun 3, 2023 Acoustic Scene Classification Data Augmentation
— Unverified 0Group channel pruning and spatial attention distilling for object detection Jun 2, 2023 Knowledge Distillation Model Compression
— Unverified 0Task-Agnostic Structured Pruning of Speech Representation Models Jun 2, 2023 Model Compression
— Unverified 0ConaCLIP: Exploring Distillation of Fully-Connected Knowledge Interaction Graph for Lightweight Text-Image Retrieval May 28, 2023 Image Retrieval Knowledge Distillation
— Unverified 02-bit Conformer quantization for automatic speech recognition May 26, 2023 Automatic Speech Recognition Automatic Speech Recognition (ASR)
— Unverified 0