SOTAVerified

Neural Network Compression

Papers

Showing 101150 of 193 papers

TitleStatusHype
Scalable Neural Network Compression and Pruning Using Hard Clustering and L1 Regularization0
SCANN: Synthesis of Compact and Accurate Neural Networks0
Semi-tensor Product-based TensorDecomposition for Neural Network Compression0
Soft-to-Hard Vector Quantization for End-to-End Learning Compressible Representations0
Sparse matrix products for neural network compression0
SPC-NeRF: Spatial Predictive Compression for Voxel Based Radiance Field0
Stabilizing Quantization-Aware Training by Implicit-Regularization on Hessian Matrix0
Survey on Computer Vision Techniques for Internet-of-Things Devices0
Taxonomy and Evaluation of Structured Compression of Convolutional Neural Networks0
The Impact of Quantization and Pruning on Deep Reinforcement Learning Models0
ThiNet: A Filter Level Pruning Method for Deep Neural Network Compression0
Tiled Bit Networks: Sub-Bit Neural Network Compression Through Reuse of Learnable Binary Vectors0
TOCO: A Framework for Compressing Neural Network Models Based on Tolerance Analysis0
DeepCABAC: Context-adaptive binary arithmetic coding for deep neural network compression0
Toward Compact Parameter Representations for Architecture-Agnostic Neural Network Compression0
Towards Explaining Deep Neural Network Compression Through a Probabilistic Latent Space0
Transform Quantization for CNN (Convolutional Neural Network) Compression0
Tropical Geometrical Zonotope Reduction as Applied to Neural Network Compression.0
TropNNC: Structured Neural Network Compression Using Tropical Geometry0
UMEC: Unified model and embedding compression for efficient recommendation systems0
Understanding the Effect of the Long Tail on Neural Network Compression0
Unified Framework for Neural Network Compression via Decomposition and Optimal Rank Selection0
Universal Deep Neural Network Compression0
VQN: Variable Quantization Noise for Neural Network Compression0
Weight Normalization based Quantization for Deep Neural Network Compression0
What is Left After Distillation? How Knowledge Transfer Impacts Fairness and Bias0
Generalized Ternary Connect: End-to-End Learning and Compression of Multiplication-Free Deep Neural Networks0
GranQ: Granular Zero-Shot Quantization with Channel-Wise Activation Scaling in QAT0
Grokking as Compression: A Nonlinear Complexity Perspective0
Guaranteed Quantization Error Computation for Neural Network Model Compression0
Hardware-Guided Symbiotic Training for Compact, Accurate, yet Execution-Efficient LSTM0
HEMP: High-order Entropy Minimization for neural network comPression0
How Informative is the Approximation Error from Tensor Decomposition for Neural Network Compression?0
Hybrid Tensor Decomposition in Neural Network Compression0
Is Quantum Optimization Ready? An Effort Towards Neural Network Compression using Adiabatic Quantum Computing0
Learning Filter Pruning Criteria for Deep Convolutional Neural Networks Acceleration0
Lightweight Attribute Localizing Models for Pedestrian Attribute Recognition0
Linearity-based neural network compression0
Compressing 3DCNNs Based on Tensor Train Decomposition0
Low-Rank Matrix Approximation for Neural Network Compression0
Minimally Invasive Surgery for Sparse Neural Networks in Contrastive Manner0
MINT: Deep Network Compression via Mutual Information-based Neuron Trimming0
Partial Binarization of Neural Networks for Budget-Aware Efficient Learning0
MLPrune: Multi-Layer Pruning for Automated Neural Network Compression0
Exact Backpropagation in Binary Weighted Networks with Group Weight TransformationsCode0
WHC: Weighted Hybrid Criterion for Filter Pruning on Convolutional Neural NetworksCode0
COP: Customized Deep Model Compression via Regularized Correlation-Based Filter-Level PruningCode0
MUSCO: Multi-Stage Compression of neural networksCode0
A Programmable Approach to Neural Network CompressionCode0
Parallel Blockwise Knowledge Distillation for Deep Neural Network CompressionCode0
Show:102550
← PrevPage 3 of 4Next →

No leaderboard results yet.