SOTAVerified

Quantization

Quantization is a promising technique to reduce the computation cost of neural network training, which can replace high-cost floating-point numbers (e.g., float32) with low-cost fixed-point numbers (e.g., int8/int16).

Source: Adaptive Precision Training: Quantify Back Propagation in Neural Networks with Fixed-point Numbers

Papers

Showing 27012750 of 4925 papers

TitleStatusHype
Low Rank Optimization for Efficient Deep Learning: Making A Balance between Compact Architecture and Fast Training0
Fighting over-fitting with quantization for learning deep neural networks on noisy labels0
Quantized Zero Dynamics Attacks against Sampled-data Control Systems0
R2 Loss: Range Restriction Loss for Model Compression and Quantization0
Rediscovering Hashed Random Projections for Efficient Quantization of Contextualized Sentence EmbeddingsCode0
Bag of Tricks with Quantized Convolutional Neural Networks for image classification0
Modular Quantization-Aware Training for 6D Object Pose EstimationCode0
Regularized Vector Quantization for Tokenized Image Synthesis0
Entropy Coding Improvement for Low-complexity Compressive Auto-encoders0
Greener yet Powerful: Taming Large Code Generation Models with Quantization0
Dynamic Stashing Quantization for Efficient Transformer Training0
QuickSRNet: Plain Single-Image Super-Resolution Architecture for Faster Inference on Mobile Platforms0
A Privacy Preserving System for Movie Recommendations Using Federated Learning0
ML Codebook Design for Initial Access and CSI Type-II Feedback in Sub-6GHz 5G NR0
MetaGrad: Adaptive Gradient Quantization with Hypernetworks0
Fixed-point quantization aware training for on-device keyword-spotting0
Summary Statistic Privacy in Data SharingCode0
Ultra-low Power Deep Learning-based Monocular Relative Localization Onboard Nano-quadrotors0
Rotation Invariant Quantization for Model CompressionCode0
Google USM: Scaling Automatic Speech Recognition Beyond 100 Languages0
Ternary Quantization: A Survey0
Boosting Distributed Full-graph GNN Training with Asynchronous One-bit Communication0
Comprehensive Comparisons of Uniform Quantization in Deep Image CompressionCode0
A Probabilistic Reformulation Technique for Discrete RIS Optimization in Wireless Systems0
Ultra-low Precision Multiplication-free Training for Deep Neural Networks0
Wireless End-to-End Image Transmission System using Semantic Communications0
JND-Based Perceptual Optimization For Learned Image Compression0
The Effect of Points Dispersion on the k-nn Search in Random Projection ForestsCode0
DyBit: Dynamic Bit-Precision Numbers for Efficient Quantized Neural Network Inference0
Teacher Intervention: Improving Convergence of Quantization Aware Training for Ultra-Low Precision TransformersCode0
Approximate spectral clustering density-based similarity for noisy datasetsCode0
HDR image watermarking using saliency detection and quantization index modulation0
Quantized Low-Rank Multivariate Regression with Random Dithering0
Refining a k-nearest neighbor graph for a computationally efficient spectral clusteringCode0
Optical Transformers0
Fixflow: A Framework to Evaluate Fixed-point Arithmetic in Light-Weight CNN Inference0
Rethinking Data-Free Quantization as a Zero-Sum GameCode0
An anatomy-based V1 model: Extraction of Low-level Features, Reduction of distortion and a V1-inspired SOM0
Evaluation of Linear Implicit Quantized State System method for analyzing mission performance of power systems0
LiQuiD-MIMO Radar: Distributed MIMO Radar with Low-Bit Quantization0
Speech Enhancement with Multi-granularity Vector Quantization0
Sparse-SignSGD with Majority Vote for Communication-Efficient Distributed Learning0
Towards Optimal Compression: Joint Pruning and Quantization0
SEAM: Searching Transferable Mixed-Precision Quantization Policy through Large Margin Regularization0
gpcgc: a green point cloud geometry coding method0
Step by Step Loss Goes Very Far: Multi-Step Quantization for Adversarial Text AttacksCode0
A Practical Mixed Precision Algorithm for Post-Training Quantization0
Feature Affinity Assisted Knowledge Distillation and Quantization of Deep Neural Networks on Label-Free Data0
Data Quality-aware Mixed-precision Quantization via Hybrid Reinforcement Learning0
Quantized Distributed Training of Large Models with Convergence Guarantees0
Show:102550
← PrevPage 55 of 99Next →

Benchmark Results

#ModelMetricClaimedVerifiedStatus
1FQ-ViT (ViT-L)Top-1 Accuracy (%)85.03Unverified
2FQ-ViT (ViT-B)Top-1 Accuracy (%)83.31Unverified
3FQ-ViT (Swin-B)Top-1 Accuracy (%)82.97Unverified
4FQ-ViT (Swin-S)Top-1 Accuracy (%)82.71Unverified
5FQ-ViT (DeiT-B)Top-1 Accuracy (%)81.2Unverified
6FQ-ViT (Swin-T)Top-1 Accuracy (%)80.51Unverified
7FQ-ViT (DeiT-S)Top-1 Accuracy (%)79.17Unverified
8Xception W8A8Top-1 Accuracy (%)78.97Unverified
9ADLIK-MO-ResNet50-W4A4Top-1 Accuracy (%)77.88Unverified
10ADLIK-MO-ResNet50-W3A4Top-1 Accuracy (%)77.34Unverified
#ModelMetricClaimedVerifiedStatus
13DCNN_VIVA_3MAP160,327.04Unverified
2DTQMAP0.79Unverified
#ModelMetricClaimedVerifiedStatus
1OutEffHop-Bert_basePerplexity6.3Unverified
2OutEffHop-Bert_basePerplexity6.21Unverified
#ModelMetricClaimedVerifiedStatus
1Accuracy98.13Unverified
#ModelMetricClaimedVerifiedStatus
1Accuracy92.92Unverified
#ModelMetricClaimedVerifiedStatus
1SSD ResNet50 V1 FPN 640x640MAP34.3Unverified
#ModelMetricClaimedVerifiedStatus
1TAR @ FAR=1e-495.13Unverified
#ModelMetricClaimedVerifiedStatus
1TAR @ FAR=1e-496.38Unverified
#ModelMetricClaimedVerifiedStatus
13DCNN_VIVA_5All84,809,664Unverified
#ModelMetricClaimedVerifiedStatus
1Accuracy99.8Unverified