Model Adaptation for Time Constrained Embodied Control Jun 17, 2024 Autonomous Driving Decision Making
— Unverified 0An Empirical Investigation of Matrix Factorization Methods for Pre-trained Transformers Jun 17, 2024 Model Compression text-classification
— Unverified 0Compress then Serve: Serving Thousands of LoRA Adapters with Little Overhead Jun 17, 2024 GPU Model Compression
— Unverified 0Knowledge Distillation in Federated Learning: a Survey on Long Lasting Challenges and New Solutions Jun 16, 2024 Federated Learning Knowledge Distillation
— Unverified 0Implicit Neural Representation for Videos Based on Residual Connection Jun 15, 2024 Image Reconstruction Model Compression
— Unverified 0EncCluster: Scalable Functional Encryption in Federated Learning through Weight Clustering and Probabilistic Filters Jun 13, 2024 Federated Learning Model Compression
— Unverified 0PC-LoRA: Low-Rank Adaptation for Progressive Model Compression with Knowledge Distillation Jun 13, 2024 Knowledge Distillation Model Compression
— Unverified 0MobileAIBench: Benchmarking LLMs and LMMs for On-Device Use Cases Jun 12, 2024 Benchmarking Model Compression
— Unverified 0DistilDoc: Knowledge Distillation for Visually-Rich Document Applications Jun 12, 2024 document-image-classification Document Image Classification
— Unverified 0On the social bias of speech self-supervised models Jun 7, 2024 Model Compression Self-Supervised Learning
— Unverified 0Slicing Mutual Information Generalization Bounds for Neural Networks Jun 6, 2024 Generalization Bounds Model Compression
Code Code Available 0Enhancing In-Context Learning Performance with just SVD-Based Weight Pruning: A Theoretical Perspective Jun 6, 2024 Generalization Bounds In-Context Learning
Code Code Available 0Reweighted Solutions for Weighted Low Rank Approximation Jun 4, 2024 feature selection Model Compression
— Unverified 0Towards Efficient Deep Spiking Neural Networks Construction with Spiking Activity based Pruning Jun 3, 2024 Model Compression Network Pruning
— Unverified 0Robust Knowledge Distillation Based on Feature Variance Against Backdoored Teacher Model Jun 1, 2024 Knowledge Distillation Model Compression
Code Code Available 0Effective Interplay between Sparsity and Quantization: From Theory to Practice May 31, 2024 Computational Efficiency Model Compression
— Unverified 0LCQ: Low-Rank Codebook based Quantization for Large Language Models May 31, 2024 Model Compression Quantization
— Unverified 0Dual sparse training framework: inducing activation map sparsity via Transformed 1 regularization May 30, 2024 Model Compression
— Unverified 0Occam Gradient Descent May 30, 2024 image-classification Image Classification
Code Code Available 0subMFL: Compatiple subModel Generation for Federated Learning in Device Heterogenous Environment May 30, 2024 Federated Learning Model Compression
Code Code Available 0NV-Embed: Improved Techniques for Training LLMs as Generalist Embedding Models May 27, 2024 Information Retrieval Language Modelling
— Unverified 0Efficient Model Compression for Hierarchical Federated Learning May 27, 2024 Edge-computing Federated Learning
— Unverified 0ExtremeMETA: High-speed Lightweight Image Segmentation Model by Remodeling Multi-channel Metamaterial Imagers May 27, 2024 Image Segmentation Model Compression
— Unverified 0Efficiency optimization of large-scale language models based on deep learning in natural language processing tasks May 20, 2024 Inference Optimization Knowledge Distillation
— Unverified 0TinyM^2Net-V3: Memory-Aware Compressed Multimodal Deep Neural Networks for Sustainable Edge Deployment May 20, 2024 Knowledge Distillation Model Compression
— Unverified 0Densely Distilling Cumulative Knowledge for Continual Learning May 16, 2024 All Continual Learning
— Unverified 0AdaKD: Dynamic Knowledge Distillation of ASR models using Adaptive Loss Weighting May 11, 2024 Knowledge Distillation Model Compression
— Unverified 0Characterizing the Accuracy -- Efficiency Trade-off of Low-rank Decomposition in Language Models May 10, 2024 AI Agent Model Compression
— Unverified 0NurtureNet: A Multi-task Video-based Approach for Newborn Anthropometry May 9, 2024 Model Compression
— Unverified 0From Algorithm to Hardware: A Survey on Efficient and Safe Deployment of Deep Neural Networks May 9, 2024 Knowledge Distillation Model Compression
— Unverified 0Light Field Compression Based on Implicit Neural Representation May 7, 2024 Model Compression
— Unverified 0Communication-Efficient Federated Learning with Adaptive Compression under Dynamic Bandwidth May 6, 2024 Federated Learning Model Compression
— Unverified 0Trio-ViT: Post-Training Quantization and Acceleration for Softmax-Free Efficient Vision Transformer May 6, 2024 Efficient ViTs Model Compression
Code Code Available 0Iterative Filter Pruning for Concatenation-based CNN Architectures May 4, 2024 Model Compression
Code Code Available 0Dependency-Aware Semi-Structured Sparsity of GLU Variants in Large Language Models May 3, 2024 Computational Efficiency Model Compression
— Unverified 0FedGreen: Carbon-aware Federated Learning with Model Size Adaptation Apr 23, 2024 Federated Learning Model Compression
— Unverified 0Rapid Deployment of DNNs for Edge Computing via Structured Pruning at Initialization Apr 22, 2024 Edge-computing Model Compression
— Unverified 0Data-free Knowledge Distillation for Fine-grained Visual Categorization Apr 18, 2024 Data-free Knowledge Distillation Fine-Grained Visual Categorization
Code Code Available 0Understanding the Performance Horizon of the Latest ML Workloads with NonGEMM Workloads Apr 17, 2024 Model Compression
— Unverified 0Comprehensive Survey of Model Compression and Speed up for Vision Transformers Apr 16, 2024 Computational Efficiency Edge-computing
— Unverified 0Structured Model Pruning for Efficient Inference in Computational Pathology Apr 12, 2024 Instance Segmentation Model Compression
— Unverified 0Simplifying Two-Stage Detectors for On-Device Inference in Remote Sensing Apr 11, 2024 Model Compression Object
— Unverified 0Bayesian Federated Model Compression for Communication and Computation Efficiency Apr 11, 2024 Bayesian Inference Federated Learning
— Unverified 0Multilingual Brain Surgeon: Large Language Models Can be Compressed Leaving No Language Behind Apr 6, 2024 Model Compression
Code Code Available 0Improve Knowledge Distillation via Label Revision and Data Selection Apr 3, 2024 Knowledge Distillation Model Compression
— Unverified 0Knowledge Distillation with Multi-granularity Mixture of Priors for Image Super-Resolution Apr 3, 2024 Image Super-Resolution Knowledge Distillation
— Unverified 0Automated Inference of Graph Transformation Rules Apr 3, 2024 Model Compression
— Unverified 0On Linearizing Structured Data in Encoder-Decoder Language Models: Insights from Text-to-SQL Apr 3, 2024 Decoder Knowledge Graphs
— Unverified 0Enhancing Inference Efficiency of Large Language Models: Investigating Optimization Strategies and Architectural Innovations Apr 2, 2024 Model Compression
— Unverified 0Instance-Aware Group Quantization for Vision Transformers Apr 1, 2024 image-classification Image Classification
— Unverified 0