Runtime Tunable Tsetlin Machines for Edge Inference on eFPGAs Feb 10, 2025 Model Compression Resynthesis
— Unverified 0Systematic Outliers in Large Language Models Feb 10, 2025 Model Compression
Code Code Available 0Synergistic Effects of Knowledge Distillation and Structured Pruning for Self-Supervised Speech Models Feb 9, 2025 Knowledge Distillation Model Compression
— Unverified 0Theoretical Guarantees for Low-Rank Compression of Deep Neural Networks Feb 4, 2025 Low-rank compression Model Compression
— Unverified 0Accelerating Linear Recurrent Neural Networks for the Edge with Unstructured Sparsity Feb 3, 2025 Audio Denoising Denoising
— Unverified 0MIND: Modality-Informed Knowledge Distillation Framework for Multimodal Clinical Prediction Tasks Feb 3, 2025 Imputation Knowledge Distillation
— Unverified 0Role of Mixup in Topological Persistence Based Knowledge Distillation for Wearable Sensor Data Feb 2, 2025 Data Augmentation Knowledge Distillation
— Unverified 0Attention Sinks and Outlier Features: A 'Catch, Tag, and Release' Mechanism for Embeddings Feb 2, 2025 Model Compression TAG
— Unverified 0Huff-LLM: End-to-End Lossless Compression for Efficient LLM Inference Feb 2, 2025 Model Compression Quantization
— Unverified 0Pivoting Factorization: A Compact Meta Low-Rank Representation of Sparsity for Efficient Inference in Large Language Models Jan 31, 2025 GPU Model Compression
— Unverified 0Efficient Supernet Training with Orthogonal Softmax for Scalable ASR Model Compression Jan 31, 2025 Model Compression
— Unverified 0Perforated Backpropagation: A Neuroscience Inspired Extension to Artificial Neural Networks Jan 29, 2025 Drug Discovery Language Modeling
Code Code Available 0TAID: Temporally Adaptive Interpolated Distillation for Efficient Knowledge Transfer in Language Models Jan 28, 2025 Knowledge Distillation Model Compression
— Unverified 0On Accelerating Edge AI: Optimizing Resource-Constrained Environments Jan 25, 2025 Knowledge Distillation Model Compression
— Unverified 0You Only Prune Once: Designing Calibration-Free Model Compression With Policy Learning Jan 25, 2025 Model Compression
— Unverified 0SwiftPrune: Hessian-Free Weight Pruning for Large Language Models Jan 24, 2025 Model Compression Quantization
— Unverified 0Practical quantum federated learning and its experimental demonstration Jan 22, 2025 Federated Learning Model Compression
— Unverified 0MultiPruner: Balanced Structure Removal in Foundation Models Jan 17, 2025 Model Compression
— Unverified 0FASP: Fast and Accurate Structured Pruning of Large Language Models Jan 16, 2025 GPU Model Compression
— Unverified 0Knowledge Distillation for Image Restoration : Simultaneous Learning from Degraded and Clean Images Jan 16, 2025 Decoder Image Reconstruction
— Unverified 0Atleus: Accelerating Transformers on the Edge Enabled by 3D Heterogeneous Manycore Architectures Jan 16, 2025 Model Compression Quantization
— Unverified 0SWSC: Shared Weight for Similar Channel in LLM Jan 15, 2025 Model Compression
— Unverified 0Tensorization of neural networks for improved privacy and interpretability Jan 10, 2025 Model Compression
Code Code Available 0Neural Architecture Codesign for Fast Physics Applications Jan 9, 2025 High-Level Synthesis Model Compression
Code Code Available 0UPAQ: A Framework for Real-Time and Energy-Efficient 3D Object Detection in Autonomous Vehicles Jan 8, 2025 3D Object Detection Autonomous Vehicles
— Unverified 0CURing Large Models: Compression via CUR Decomposition Jan 8, 2025 Model Compression
— Unverified 0Effective and Efficient Mixed Precision Quantization of Speech Foundation Models Jan 7, 2025 Model Compression parameter estimation
— Unverified 0Strategic Fusion Optimizes Transformer Compression Jan 5, 2025 Knowledge Distillation Model Compression
— Unverified 0Optimizing Small Language Models for In-Vehicle Function-Calling Jan 4, 2025 Model Compression Quantization
— Unverified 0DeepCompress-ViT: Rethinking Model Compression to Enhance Efficiency of Vision Transformers at the Edge Jan 1, 2025 Model Compression
Code Code Available 0Once-Tuning-Multiple-Variants: Tuning Once and Expanded as Multiple Vision-Language Model Variants Jan 1, 2025 Language Modeling Language Modelling
— Unverified 0Random Conditioning for Diffusion Model Compression with Distillation Jan 1, 2025 Denoising Knowledge Distillation
— Unverified 0Improving Acoustic Scene Classification in Low-Resource Conditions Dec 30, 2024 Acoustic Scene Classification Classification
— Unverified 0Feature Alignment-Based Knowledge Distillation for Efficient Compression of Large Language Models Dec 27, 2024 Knowledge Distillation Model Compression
— Unverified 0Optimization and Scalability of Collaborative Filtering Algorithms in Large Language Models Dec 25, 2024 Collaborative Filtering Computational Efficiency
— Unverified 0HTR-JAND: Handwritten Text Recognition with Joint Attention Network and Knowledge Distillation Dec 24, 2024 Computational Efficiency Handwritten Text Recognition
Code Code Available 0Edge-AI for Agriculture: Lightweight Vision Models for Disease Detection in Resource-Limited Settings Dec 23, 2024 Computational Efficiency Federated Learning
— Unverified 0GQSA: Group Quantization and Sparsity for Accelerating Large Language Model Inference Dec 23, 2024 GPU Language Modeling
— Unverified 0CoSurfGS:Collaborative 3D Surface Gaussian Splatting with Distributed Learning for Large Scene Reconstruction Dec 23, 2024 3DGS GPU
— Unverified 0Singular Value Scaling: Efficient Generative Model Compression via Pruned Weights Refinement Dec 23, 2024 Model Compression
Code Code Available 0Lightweight Design and Optimization methods for DCNNs: Progress and Futures Dec 22, 2024 Autonomous Driving Deep Learning
— Unverified 0Semantics Prompting Data-Free Quantization for Low-Bit Vision Transformers Dec 21, 2024 Data Free Quantization Model Compression
— Unverified 0Deploying Foundation Model Powered Agent Services: A Survey Dec 18, 2024 model Model Compression
— Unverified 0RemoteTrimmer: Adaptive Structural Pruning for Remote Sensing Image Classification Dec 17, 2024 Classification image-classification
Code Code Available 0TrimLLM: Progressive Layer Dropping for Domain-Specific LLMs Dec 15, 2024 Model Compression Quantization
— Unverified 0Can Students Beyond The Teacher? Distilling Knowledge from Teacher's Bias Dec 13, 2024 Knowledge Distillation Model Compression
— Unverified 0Activation Sparsity Opportunities for Compressing General Large Language Models Dec 13, 2024 Model Compression
— Unverified 0Optimising TinyML with Quantization and Distillation of Transformer and Mamba Models for Indoor Localisation on Edge Devices Dec 12, 2024 Knowledge Distillation Mamba
— Unverified 0Low-Rank Correction for Quantized LLMs Dec 10, 2024 Model Compression Quantization
— Unverified 0Lossless Model Compression via Joint Low-Rank Factorization Optimization Dec 9, 2024 Model Compression Model Optimization
— Unverified 0