Analysis of memory consumption by neural networks based on hyperparameters Oct 21, 2021 Deep Learning Model Compression
— Unverified 00 Neural Regularized Domain Adaptation for Chinese Word Segmentation Dec 1, 2017 Chinese Word Segmentation Domain Adaptation
— Unverified 00 NeuSemSlice: Towards Effective DNN Model Maintenance via Neuron-level Semantic Slicing Jul 26, 2024 Model Compression Semantic Similarity
— Unverified 00 Noisy Neural Network Compression for Analog Storage Devices Oct 19, 2020 Knowledge Distillation Model Compression
— Unverified 00 Understanding the Performance Horizon of the Latest ML Workloads with NonGEMM Workloads Apr 17, 2024 Model Compression
— Unverified 00 Non-Structured DNN Weight Pruning -- Is It Beneficial in Any Platform? Jul 3, 2019 Model Compression Quantization
— Unverified 00 Normalized Feature Distillation for Semantic Segmentation Jul 12, 2022 Knowledge Distillation Model Compression
— Unverified 00 Norm Tweaking: High-performance Low-bit Quantization of Large Language Models Sep 6, 2023 Model Compression Quantization
— Unverified 00 NurtureNet: A Multi-task Video-based Approach for Newborn Anthropometry May 9, 2024 Model Compression
— Unverified 00 NV-Embed: Improved Techniques for Training LLMs as Generalist Embedding Models May 27, 2024 Information Retrieval Language Modelling
— Unverified 00 NVRC: Neural Video Representation Compression Sep 11, 2024 Model Compression Quantization
— Unverified 00 oBERTa: Improving Sparse Transfer Learning via improved initialization, distillation, and pruning regimes Mar 30, 2023 Knowledge Distillation Model Compression
— Unverified 00 Towards efficient deep autoencoders for multivariate time series anomaly detection Mar 4, 2024 Anomaly Detection Model Compression
— Unverified 00 An Algorithm-Hardware Co-Optimized Framework for Accelerating N:M Sparse Transformers Aug 12, 2022 Computational Efficiency Model Compression
— Unverified 00 On Accelerating Edge AI: Optimizing Resource-Constrained Environments Jan 25, 2025 Knowledge Distillation Model Compression
— Unverified 00 On Achieving Privacy-Preserving State-of-the-Art Edge Intelligence Feb 10, 2023 Edge-computing Model Compression
— Unverified 00 Data-Independent Neural Pruning via Coresets Jul 9, 2019 Model Compression Network Pruning
— Unverified 00 On Attention Redundancy: A Comprehensive Study Jun 1, 2021 Model Compression Sentence
— Unverified 00 Onboard Optimization and Learning: A Survey May 7, 2025 Decision Making Model Compression
— Unverified 00 Once-Tuning-Multiple-Variants: Tuning Once and Expanded as Multiple Vision-Language Model Variants Jan 1, 2025 Language Modeling Language Modelling
— Unverified 00 On-Device Document Classification using multimodal features Jan 6, 2021 Classification Document Classification
— Unverified 00 An Efficient Real-Time Object Detection Framework on Resource-Constricted Hardware Devices via Software and Hardware Co-design Aug 2, 2024 Model Compression Neural Network Compression
— Unverified 00 Towards Efficient Deep Spiking Neural Networks Construction with Spiking Activity based Pruning Jun 3, 2024 Model Compression Network Pruning
— Unverified 00 On-Device Qwen2.5: Efficient LLM Inference with Model Compression and Hardware Acceleration Apr 24, 2025 CPU Model Compression
— Unverified 00 One-Shot Model for Mixed-Precision Quantization Jan 1, 2023 model Model Compression
— Unverified 00 One Teacher is Enough? Pre-trained Language Model Distillation from Multiple Teachers Jun 2, 2021 Knowledge Distillation Language Modeling
— Unverified 00 One Weight Bitwidth to Rule Them All Aug 22, 2020 All image-classification
— Unverified 00 On Linearizing Structured Data in Encoder-Decoder Language Models: Insights from Text-to-SQL Apr 3, 2024 Decoder Knowledge Graphs
— Unverified 00 Online Cross-Layer Knowledge Distillation on Graph Neural Networks with Deep Supervision Oct 25, 2022 Knowledge Distillation Model Compression
— Unverified 00 Towards Efficient Full 8-bit Integer DNN Online Training on Resource-limited Devices without Batch Normalization May 27, 2021 Model Compression Quantization
— Unverified 00 A Model Compression Method with Matrix Product Operators for Speech Enhancement Oct 10, 2020 Model Compression Speech Enhancement
— Unverified 00 Online Model Compression for Federated Learning with Large Models May 6, 2022 Federated Learning Model Compression
— Unverified 00 On Multilingual Encoder Language Model Compression for Low-Resource Languages May 22, 2025 Knowledge Distillation Language Modeling
— Unverified 00 On the Adversarial Robustness of Quantized Neural Networks May 1, 2021 Adversarial Robustness Model Compression
— Unverified 00 On the Compression of Recurrent Neural Networks with an Application to LVCSR acoustic modeling for Embedded Speech Recognition Mar 25, 2016 Model Compression speech-recognition
— Unverified 00 On the Demystification of Knowledge Distillation: A Residual Network Perspective Jun 30, 2020 Knowledge Distillation Model Compression
— Unverified 00 Towards Efficient Tensor Decomposition-Based DNN Model Compression with Optimization Framework Jul 26, 2021 image-classification Image Classification
— Unverified 00 On the Effectiveness of Low-Rank Matrix Factorization for LSTM Model Compression Aug 27, 2019 Model Compression
— Unverified 00 On the Impact of Quantization and Pruning of Self-Supervised Speech Models for Downstream Speech Recognition Tasks "In-the-Wild'' Sep 25, 2023 Data Augmentation Model Compression
— Unverified 00 On the social bias of speech self-supervised models Jun 7, 2024 Model Compression Self-Supervised Learning
— Unverified 00 Weight Squeezing: Reparameterization for Knowledge Transfer and Model Compression Oct 14, 2020 General Classification Model Compression
— Unverified 00 A Mixed Integer Programming Approach for Verifying Properties of Binarized Neural Networks Mar 11, 2022 Collision Avoidance Model Compression
— Unverified 00 Optimal Policy Sparsification and Low Rank Decomposition for Deep Reinforcement Learning Mar 10, 2024 Deep Reinforcement Learning Edge-computing
— Unverified 00 Optimising TinyML with Quantization and Distillation of Transformer and Mamba Models for Indoor Localisation on Edge Devices Dec 12, 2024 Knowledge Distillation Mamba
— Unverified 00 Optimization and Scalability of Collaborative Filtering Algorithms in Large Language Models Dec 25, 2024 Collaborative Filtering Computational Efficiency
— Unverified 00 Optimize Deep Convolutional Neural Network with Ternarized Weights and High Accuracy Jul 20, 2018 Model Compression Vocal Bursts Intensity Prediction
— Unverified 00 Optimizing LLMs for Resource-Constrained Environments: A Survey of Model Compression Techniques May 5, 2025 Knowledge Distillation Mixture-of-Experts
— Unverified 00 Optimizing Singular Spectrum for Large Language Model Compression Feb 20, 2025 Language Modeling Language Modelling
— Unverified 00 Optimizing Small Language Models for In-Vehicle Function-Calling Jan 4, 2025 Model Compression Quantization
— Unverified 00 Optimizing Traffic Signal Control using High-Dimensional State Representation and Efficient Deep Reinforcement Learning Nov 12, 2024 Deep Reinforcement Learning Model Compression
— Unverified 00