SOTAVerified

Neural Network Compression

Papers

Showing 150 of 193 papers

TitleStatusHype
Linearity-based neural network compression0
MUC-G4: Minimal Unsat Core-Guided Incremental Verification for Deep Neural Network Compression0
Is Quantum Optimization Ready? An Effort Towards Neural Network Compression using Adiabatic Quantum Computing0
Certified Neural Approximations of Nonlinear DynamicsCode0
Low-Rank Matrix Approximation for Neural Network Compression0
GranQ: Granular Zero-Shot Quantization with Channel-Wise Activation Scaling in QAT0
Stabilizing Quantization-Aware Training by Implicit-Regularization on Hessian Matrix0
Compression of Site-Specific Deep Neural Networks for Massive MIMO Precoding0
A Novel Structure-Agnostic Multi-Objective Approach for Weight-Sharing Compression in Deep Neural Networks0
What is Left After Distillation? How Knowledge Transfer Impacts Fairness and Bias0
Efficient and Robust Knowledge Distillation from A Stronger Teacher Based on Correlation Matching0
Language Models as Zero-shot Lossless Gradient Compressors: Towards General Neural Parameter Prior ModelsCode0
Adaptive Error-Bounded Hierarchical Matrices for Efficient Neural Network Compression0
TropNNC: Structured Neural Network Compression Using Tropical Geometry0
Unified Framework for Neural Network Compression via Decomposition and Optimal Rank Selection0
Convolutional Neural Network Compression Based on Low-Rank Decomposition0
Condensed Sample-Guided Model Inversion for Knowledge Distillation0
An Efficient Real-Time Object Detection Framework on Resource-Constricted Hardware Devices via Software and Hardware Co-design0
Tiled Bit Networks: Sub-Bit Neural Network Compression Through Reuse of Learnable Binary Vectors0
The Impact of Quantization and Pruning on Deep Reinforcement Learning Models0
Neural Network Compression for Reinforcement Learning Tasks0
Torch2Chip: An End-to-end Customizable Deep Neural Network Compression and Deployment Toolkit for Prototype Hardware Accelerator DesignCode2
Towards Explaining Deep Neural Network Compression Through a Probabilistic Latent Space0
SPC-NeRF: Spatial Predictive Compression for Voxel Based Radiance Field0
Towards Meta-Pruning via Optimal TransportCode1
EPSD: Early Pruning with Self-Distillation for Efficient Model Compression0
Convolutional Neural Network Compression via Dynamic Parameter Rank Pruning0
Balanced and Deterministic Weight-sharing Helps Network Performance0
ABKD: Graph Neural Network Compression with Attention-Based Knowledge Distillation0
Grokking as Compression: A Nonlinear Complexity Perspective0
Causal-DFQ: Causality Guided Data-free Network QuantizationCode0
A Survey on Deep Neural Network Pruning-Taxonomy, Comparison, Analysis, and RecommendationsCode2
Quantization Aware Factorization for Deep Neural Network Compression0
Survey on Computer Vision Techniques for Internet-of-Things Devices0
Model Compression Methods for YOLOv5: A Review0
Lightweight Attribute Localizing Models for Pedestrian Attribute Recognition0
Neural Network Compression using Binarization and Few Full-Precision Weights0
Implicit Compressibility of Overparametrized Neural Networks Trained with Heavy-Tailed SGDCode0
Understanding the Effect of the Long Tail on Neural Network Compression0
End-to-End Neural Network Compression via _1_2 Regularized Latency Surrogates0
Modular Transformers: Compressing Transformers into Modularized Layers for Flexible Efficient Inference0
Variation Spaces for Multi-Output Neural Networks: Insights on Multi-Task Learning and Network CompressionCode0
Evaluation Metrics for DNNs Compression0
How Informative is the Approximation Error from Tensor Decomposition for Neural Network Compression?0
Guaranteed Quantization Error Computation for Neural Network Model Compression0
SwiftTron: An Efficient Hardware Accelerator for Quantized TransformersCode1
WHC: Weighted Hybrid Criterion for Filter Pruning on Convolutional Neural NetworksCode0
DepGraph: Towards Any Structural PruningCode4
Magnitude and Similarity based Variable Rate Filter Pruning for Efficient Convolution Neural NetworksCode0
PD-Quant: Post-Training Quantization based on Prediction Difference MetricCode1
Show:102550
← PrevPage 1 of 4Next →

No leaderboard results yet.