Privacy-Preserving SAM Quantization for Efficient Edge Intelligence in Healthcare Sep 14, 2024 Data Free Quantization Image Segmentation
— Unverified 00 Private Model Compression via Knowledge Distillation Nov 13, 2018 Knowledge Distillation model
— Unverified 00 Train Flat, Then Compress: Sharpness-Aware Minimization Learns More Compressible Models May 25, 2022 Model Compression Quantization
— Unverified 00 Inferring ECG from PPG for Continuous Cardiac Monitoring Using Lightweight Neural Network Dec 9, 2020 Model Compression
— Unverified 00 Progressive Weight Pruning of Deep Neural Networks using ADMM Oct 17, 2018 Edge-computing Model Compression
— Unverified 00 Pro-KD: Progressive Distillation by Following the Footsteps of the Teacher Oct 16, 2021 image-classification Image Classification
— Unverified 00 ALF: Autoencoder-based Low-rank Filter-sharing for Efficient Convolutional Neural Networks Jul 27, 2020 Model Compression
— Unverified 00 âLearning-Compressionâ Algorithms for Neural Net Pruning Jun 1, 2018 Model Compression Network Pruning
— Unverified 00 Prototype-based Personalized Pruning Mar 25, 2021 Keyword Spotting Model Compression
— Unverified 00 Prototypical Contrastive Predictive Coding Sep 29, 2021 Contrastive Learning Knowledge Distillation
— Unverified 00 Provable Benefits of Overparameterization in Model Compression: From Double Descent to Pruning Neural Networks Dec 16, 2020 Model Compression
— Unverified 00 Training Acceleration of Low-Rank Decomposed Networks using Sequential Freezing and Rank Quantization Sep 7, 2023 Model Compression Quantization
— Unverified 00 Structured Pruning of a BERT-based Question Answering Model Oct 14, 2019 Model Compression Natural Questions
— Unverified 00 Pruning Algorithms to Accelerate Convolutional Neural Networks for Edge Applications: A Survey May 8, 2020 Model Compression Survey
— Unverified 00 Pruning at a Glance: A Structured Class-Blind Pruning Technique for Model Compression Oct 20, 2018 Model Compression
— Unverified 00 Pruning at a Glance: Global Neural Pruning for Model Compression Nov 30, 2019 Model Compression Sensitivity
— Unverified 00 What is Left After Distillation? How Knowledge Transfer Impacts Fairness and Bias Oct 10, 2024 Age/Unbiased Fairness
— Unverified 00 A Half-Space Stochastic Projected Gradient Method for Group Sparsity Regularization Jan 1, 2021 compressed sensing feature selection
— Unverified 00 What is Lost in Knowledge Distillation? Nov 7, 2023 Knowledge Distillation Model Compression
— Unverified 00 Pruning Large Language Models via Accuracy Predictor Sep 18, 2023 MMLU Model Compression
— Unverified 00 Aggressive Post-Training Compression on Extremely Large Language Models Sep 30, 2024 Model Compression Network Pruning
— Unverified 00 Pruning Ternary Quantization Jul 23, 2021 image-classification Image Classification
— Unverified 00 AfroXLMR-Comet: Multilingual Knowledge Distillation with Attention Matching for Low-Resource languages Feb 25, 2025 Knowledge Distillation Language Modeling
— Unverified 00 AACP: Model Compression by Accurate and Automatic Channel Pruning Jan 31, 2021 Model Compression Neural Architecture Search
— Unverified 00 A flexible, extensible software framework for model compression based on the LC algorithm May 15, 2020 BIG-bench Machine Learning Low-rank compression
— Unverified 00 Puppet-CNN: Input-Adaptive Convolutional Neural Networks with Model Compression using Ordinary Differential Equation Nov 19, 2024 Model Compression
— Unverified 00 PURSUhInT: In Search of Informative Hint Points Based on Layer Clustering for Knowledge Distillation Feb 26, 2021 Clustering Knowledge Distillation
— Unverified 00 Aerial Image Classification in Scarce and Unconstrained Environments via Conformal Prediction Apr 24, 2025 Conformal Prediction image-classification
— Unverified 00 QAPPA: Quantization-Aware Power, Performance, and Area Modeling of DNN Accelerators May 17, 2022 Model Compression Quantization
— Unverified 00 QD-BEV : Quantization-aware View-guided Distillation for Multi-view 3D Object Detection Aug 21, 2023 3D Object Detection Model Compression
— Unverified 00 Adversarially Robust and Explainable Model Compression with On-Device Personalization for Text Classification Jan 10, 2021 Adversarial Robustness General Classification
— Unverified 00 NPAS: A Compiler-aware Framework of Unified Network Pruning and Architecture Search for Beyond Real-Time Mobile Acceleration Dec 1, 2020 Bayesian Optimization Code Generation
— Unverified 00 Q-MambaIR: Accurate Quantized Mamba for Efficient Image Restoration Mar 27, 2025 Computational Efficiency Image Restoration
— Unverified 00 Adversarial Attacks on Machine Learning in Embedded and IoT Platforms Mar 3, 2023 Adversarial Robustness Model Compression
— Unverified 00 QTI Submission to DCASE 2021: residual normalization for device-imbalanced acoustic scene classification with efficient design Jun 28, 2022 Acoustic Scene Classification Knowledge Distillation
— Unverified 00 T-RECX: Tiny-Resource Efficient Convolutional neural networks with early-eXit Jul 14, 2022 image-classification Image Classification
— Unverified 00 Quantizing YOLOv7: A Comprehensive Study Jul 6, 2024 Model Compression object-detection
— Unverified 00 Quantum Neural Network Compression Jul 4, 2022 Model Compression Neural Network Compression
— Unverified 00 Advancing IIoT with Over-the-Air Federated Learning: The Role of Iterative Magnitude Pruning Mar 21, 2024 Federated Learning Model Compression
— Unverified 00 QuickNet: Maximizing Efficiency and Efficacy in Deep Architectures Jan 9, 2017 Model Compression
— Unverified 00 QUIDAM: A Framework for Quantization-Aware DNN Accelerator and Model Co-Exploration Jun 30, 2022 Model Compression Quantization
— Unverified 00 Quiver neural networks Jul 26, 2022 Model Compression
— Unverified 00 QuPeD: Quantized Personalization via Distillation with Applications to Federated Learning Jul 29, 2021 Federated Learning Knowledge Distillation
— Unverified 00 R2 Loss: Range Restriction Loss for Model Compression and Quantization Mar 14, 2023 Classification Model Compression
— Unverified 00 RADIN: Souping on a Budget Jan 31, 2024 Ensemble Learning Model Compression
— Unverified 00 Radio: Rate-Distortion Optimization for Large Language Model Compression May 5, 2025 Language Modeling Language Modelling
— Unverified 00 Random Conditioning for Diffusion Model Compression with Distillation Jan 1, 2025 Denoising Knowledge Distillation
— Unverified 00 Random Conditioning with Distillation for Data-Efficient Diffusion Model Compression Apr 2, 2025 Denoising Knowledge Distillation
— Unverified 00 Random Offset Block Embedding Array (ROBE) for CriteoTB Benchmark MLPerf DLRM Model : 1000 Compression and 3.1 Faster Inference Aug 4, 2021 GPU Model Compression
— Unverified 00 RAND: Robustness Aware Norm Decay For Quantized Seq2seq Models May 24, 2023 Machine Translation Model Compression
— Unverified 00