SOTAVerified

Quantization

Quantization is a promising technique to reduce the computation cost of neural network training, which can replace high-cost floating-point numbers (e.g., float32) with low-cost fixed-point numbers (e.g., int8/int16).

Source: Adaptive Precision Training: Quantify Back Propagation in Neural Networks with Fixed-point Numbers

Papers

Showing 9511000 of 4925 papers

TitleStatusHype
Model compression via distillation and quantizationCode0
Mitigating the Impact of Outlier Channels for Language Model Quantization with Activation RegularizationCode0
Mitigating Quantization Errors Due to Activation Spikes in GLU-Based LLMsCode0
Mixed Non-linear Quantization for Vision TransformersCode0
Minimize Quantization Output Error with Bias CompensationCode0
MINT: Multiplier-less INTeger Quantization for Energy Efficient Spiking Neural NetworksCode0
Minimal Random Code Learning: Getting Bits Back from Compressed Model ParametersCode0
Mirror Descent View for Neural Network QuantizationCode0
A Simple Low-bit Quantization Framework for Video Snapshot Compressive ImagingCode0
Model Compression with Adversarial Robustness: A Unified Optimization FrameworkCode0
Memory-Driven Mixed Low Precision Quantization For Enabling Deep Network Inference On MicrocontrollersCode0
Adversarial Fine-tuning of Compressed Neural Networks for Joint Improvement of Robustness and EfficiencyCode0
Merge-Friendly Post-Training Quantization for Multi-Target Domain AdaptationCode0
Continual Learning for Generative Retrieval over Dynamic CorporaCode0
MetaAug: Meta-Data Augmentation for Post-Training QuantizationCode0
Compressing Word Embeddings via Deep Compositional Code LearningCode0
McQueen : Mixed Precision Quantization of Early Exit NetworksCode0
Megapixel Image Generation with Step-Unrolled Denoising AutoencodersCode0
Compressing Vision Transformers for Low-Resource Visual LearningCode0
Maestro: Uncovering Low-Rank Structures via Trainable DecompositionCode0
Machine Learning at the Wireless Edge: Distributed Stochastic Gradient Descent Over-the-AirCode0
Continuous-variable neural-network quantum states and the quantum rotor modelCode0
Advanced Knowledge Transfer: Refined Feature Distillation for Zero-Shot Quantization in Edge ComputingCode0
LVPNet: A Latent-variable-based Prediction-driven End-to-end Framework for Lossless Compression of Medical ImagesCode0
Compressing 3D Gaussian Splatting by Noise-Substituted Vector QuantizationCode0
LVLM-Compress-Bench: Benchmarking the Broader Impact of Large Vision-Language Model CompressionCode0
NIRVANA: Neural Implicit Representations of Videos with Adaptive Networks and Autoregressive Patch-wise ModelingCode0
NITRO-D: Native Integer-only Training of Deep Convolutional Neural NetworksCode0
Make RepVGG Greater Again: A Quantization-aware ApproachCode0
LQ-Nets: Learned Quantization for Highly Accurate and Compact Deep Neural NetworksCode0
LRQ: Optimizing Post-Training Quantization for Large Language Models by Learning Low-Rank Weight-Scaling MatricesCode0
Convert, compress, correct: Three steps toward communication-efficient DNN trainingCode0
Compressed Object DetectionCode0
Low-Precision Stochastic Gradient Langevin DynamicsCode0
Fast Point Cloud Geometry Compression with Context-based Residual Coding and INR-based RefinementCode0
Compressed 3D Gaussian Splatting for Accelerated Novel View SynthesisCode0
LSQ++: Lower running time and higher recall in multi-codebook quantizationCode0
Low-bit Model Quantization for Deep Neural Networks: A SurveyCode0
Convolutional Neural Networks to Enhance Coded SpeechCode0
CoopNet: Cooperative Convolutional Neural Network for Low-Power MCUsCode0
LoTA-QAF: Lossless Ternary Adaptation for Quantization-Aware Fine-TuningCode0
Low-bit Quantization for Deep Graph Neural Networks with Smoothness-aware Message PropagationCode0
Comprehensive SNN Compression Using ADMM Optimization and Activity RegularizationCode0
Low-bit Quantization of Neural Networks for Efficient InferenceCode0
Comprehensive Comparisons of Uniform Quantization in Deep Image CompressionCode0
Comprehensive Analysis of the Object Detection Pipeline on UAVsCode0
Bit Error Robustness for Energy-Efficient DNN AcceleratorsCode0
On Quantizing Neural Representation for Variable-Rate Video CodingCode0
Compositional Sketch SearchCode0
Composite QuantizationCode0
Show:102550
← PrevPage 20 of 99Next →

Benchmark Results

#ModelMetricClaimedVerifiedStatus
1FQ-ViT (ViT-L)Top-1 Accuracy (%)85.03Unverified
2FQ-ViT (ViT-B)Top-1 Accuracy (%)83.31Unverified
3FQ-ViT (Swin-B)Top-1 Accuracy (%)82.97Unverified
4FQ-ViT (Swin-S)Top-1 Accuracy (%)82.71Unverified
5FQ-ViT (DeiT-B)Top-1 Accuracy (%)81.2Unverified
6FQ-ViT (Swin-T)Top-1 Accuracy (%)80.51Unverified
7FQ-ViT (DeiT-S)Top-1 Accuracy (%)79.17Unverified
8Xception W8A8Top-1 Accuracy (%)78.97Unverified
9ADLIK-MO-ResNet50-W4A4Top-1 Accuracy (%)77.88Unverified
10ADLIK-MO-ResNet50-W3A4Top-1 Accuracy (%)77.34Unverified
#ModelMetricClaimedVerifiedStatus
13DCNN_VIVA_3MAP160,327.04Unverified
2DTQMAP0.79Unverified
#ModelMetricClaimedVerifiedStatus
1OutEffHop-Bert_basePerplexity6.3Unverified
2OutEffHop-Bert_basePerplexity6.21Unverified
#ModelMetricClaimedVerifiedStatus
1Accuracy98.13Unverified
#ModelMetricClaimedVerifiedStatus
1Accuracy92.92Unverified
#ModelMetricClaimedVerifiedStatus
1SSD ResNet50 V1 FPN 640x640MAP34.3Unverified
#ModelMetricClaimedVerifiedStatus
1TAR @ FAR=1e-495.13Unverified
#ModelMetricClaimedVerifiedStatus
1TAR @ FAR=1e-496.38Unverified
#ModelMetricClaimedVerifiedStatus
13DCNN_VIVA_5All84,809,664Unverified
#ModelMetricClaimedVerifiedStatus
1Accuracy99.8Unverified