OPTISHEAR: Towards Efficient and Adaptive Pruning of Large Language Models via Evolutionary Optimization Feb 15, 2025 Model Compression
— Unverified 00 Oracle Teacher: Leveraging Target Information for Better Knowledge Distillation of CTC Models Nov 5, 2021 Knowledge Distillation Machine Translation
— Unverified 00 A Memory-Efficient Learning Framework for SymbolLevel Precoding with Quantized NN Weights Oct 13, 2021 Model Compression Quantization
— Unverified 00 OTOV2: Automatic, Generic, User-Friendly Mar 13, 2023 Model Compression
— Unverified 00 Outsourcing Training without Uploading Data via Efficient Collaborative Open-Source Sampling Oct 23, 2022 Model Compression
— Unverified 00 Towards Higher Ranks via Adversarial Weight Pruning Nov 29, 2023 Model Compression Network Pruning
— Unverified 00 Pacemaker: Intermediate Teacher Knowledge Distillation For On-The-Fly Convolutional Neural Network Mar 9, 2020 Knowledge Distillation Model Compression
— Unverified 00 Pangu Light: Weight Re-Initialization for Pruning and Accelerating LLMs May 26, 2025 Model Compression
— Unverified 00 Parameter Compression of Recurrent Neural Networks and Degradation of Short-term Memory Dec 2, 2016 Memorization Model Compression
— Unverified 00 AMD: Automatic Multi-step Distillation of Large-scale Vision Models Jul 5, 2024 image-classification Image Classification
— Unverified 00 Single-path Bit Sharing for Automatic Loss-aware Model Compression Jan 13, 2021 Model Compression Network Pruning
— Unverified 00 Partitioning-Guided K-Means: Extreme Empty Cluster Resolution for Extreme Model Compression Jun 24, 2023 Model Compression Quantization
— Unverified 00 AMD: Adaptive Masked Distillation for Object Detection Jan 31, 2023 Knowledge Distillation Model Compression
— Unverified 00 PatDNN: Achieving Real-Time DNN Execution on Mobile Devices with Pattern-based Weight Pruning Jan 1, 2020 Code Generation Model Compression
— Unverified 00 Towards Modality Transferable Visual Information Representation with Optimal Model Compression Aug 13, 2020 Model Compression Philosophy
— Unverified 00 AlphaTuning: Quantization-Aware Parameter-Efficient Adaptation of Large-Scale Pre-Trained Language Models Oct 8, 2022 Language Modeling Language Modelling
— Unverified 00 PCEE-BERT: Accelerating BERT Inference via Patient and Confident Early Exiting Jan 16, 2022 Model Compression
— Unverified 00 Towards Optimal Compression: Joint Pruning and Quantization Feb 15, 2023 Model Compression Neural Architecture Search
— Unverified 00 PC-LoRA: Low-Rank Adaptation for Progressive Model Compression with Knowledge Distillation Jun 13, 2024 Knowledge Distillation Model Compression
— Unverified 00 PCNN: Pattern-based Fine-Grained Regular Pruning towards Optimizing CNN Accelerators Feb 11, 2020 Model Compression
— Unverified 00 PCONV: The Missing but Desirable Sparsity in DNN Weight Pruning for Real-time Execution on Mobile Devices Sep 6, 2019 Model Compression
— Unverified 00 Pea-KD: Parameter-efficient and Accurate Knowledge Distillation on BERT Sep 30, 2020 Knowledge Distillation Model Compression
— Unverified 00 Pea-KD: Parameter-efficient and accurate Knowledge Distillation Sep 28, 2020 Knowledge Distillation Model Compression
— Unverified 00 Weight Squeezing: Reparameterization for Compression and Fast Inference May 30, 2020 Knowledge Distillation Model Compression
— Unverified 00 Penrose Tiled Low-Rank Compression and Section-Wise Q&A Fine-Tuning: A General Framework for Domain-Specific Large Language Model Adaptation Mar 28, 2025 Language Modeling Language Modelling
— Unverified 00 Towards Superior Quantization Accuracy: A Layer-sensitive Approach Mar 9, 2025 Logical Reasoning Model Compression
— Unverified 00 A Low-Power Streaming Speech Enhancement Accelerator For Edge Devices Mar 27, 2025 Model Compression Speech Enhancement
— Unverified 00 Performance Aware Convolutional Neural Network Channel Pruning for Embedded GPUs Feb 20, 2020 Model Compression Network Pruning
— Unverified 00 PERMDNN: Efficient Compressed DNN Architecture with Permuted Diagonal Matrices Apr 23, 2020 Model Compression
— Unverified 00 Perturbation of Deep Autoencoder Weights for Model Compression and Classification of Tabular Data May 17, 2022 BIG-bench Machine Learning Classification
— Unverified 00 PFGDF: Pruning Filter via Gaussian Distribution Feature for Deep Neural Networks Acceleration Jun 23, 2020 Model Compression
— Unverified 00 Weight Squeezing: Reparameterization for Knowledge Transfer and Model Compression Nov 16, 2021 Model Compression text-classification
— Unverified 00 Pivoting Factorization: A Compact Meta Low-Rank Representation of Sparsity for Efficient Inference in Large Language Models Jan 31, 2025 GPU Model Compression
— Unverified 00 A Low Effort Approach to Structured CNN Design Using PCA Dec 15, 2018 Dimensionality Reduction Model Compression
— Unverified 00 Do we need Label Regularization to Fine-tune Pre-trained Language Models? May 25, 2022 Knowledge Distillation Model Compression
— Unverified 00 Tensor Train Low-rank Approximation (TT-LoRA): Democratizing AI with Accelerated LLMs Aug 2, 2024 Machine Translation Model Compression
— Unverified 00 A Lottery Ticket Hypothesis Framework for Low-Complexity Device-Robust Neural Acoustic Scene Classification Jul 3, 2021 Acoustic Scene Classification Data Augmentation
— Unverified 00 Towards Zero-Shot Knowledge Distillation for Natural Language Processing Dec 31, 2020 Knowledge Distillation Model Compression
— Unverified 00 Robustness Challenges in Model Distillation and Pruning for Natural Language Understanding Oct 16, 2021 Knowledge Distillation Model Compression
— Unverified 00 Position-Aware Depth Decay Decoding (D^3): Boosting Large Language Model Inference Efficiency Mar 11, 2025 GSM8K Language Modeling
— Unverified 00 Aligned Weight Regularizers for Pruning Pretrained Neural Networks Apr 4, 2022 Language Modelling Model Compression
— Unverified 00 Post-Training Quantization for Video Matting Jun 12, 2025 Image Matting Model Compression
— Unverified 00 Shedding the Bits: Pushing the Boundaries of Quantization with Minifloats on FPGAs Nov 21, 2023 Model Compression Quantization
— Unverified 00 Post-Training Weighted Quantization of Neural Networks for Language Models Jan 1, 2021 Model Compression Quantization
— Unverified 00 PQK: Model Compression via Pruning, Quantization, and Knowledge Distillation Jun 25, 2021 Keyword Spotting Knowledge Distillation
— Unverified 00 Practical quantum federated learning and its experimental demonstration Jan 22, 2025 Federated Learning Model Compression
— Unverified 00 Precise Box Score: Extract More Information from Datasets to Improve the Performance of Face Detection Apr 28, 2018 Face Detection Model Compression
— Unverified 00 What do larger image classifiers memorise? Oct 9, 2023 image-classification Image Classification
— Unverified 00 Preventing Catastrophic Forgetting and Distribution Mismatch in Knowledge Distillation via Synthetic Data Aug 11, 2021 Knowledge Distillation Model Compression
— Unverified 00 Preview-based Category Contrastive Learning for Knowledge Distillation Oct 18, 2024 Contrastive Learning Knowledge Distillation
— Unverified 00