SOTAVerified

Quantization

Quantization is a promising technique to reduce the computation cost of neural network training, which can replace high-cost floating-point numbers (e.g., float32) with low-cost fixed-point numbers (e.g., int8/int16).

Source: Adaptive Precision Training: Quantify Back Propagation in Neural Networks with Fixed-point Numbers

Papers

Showing 10011025 of 4925 papers

TitleStatusHype
How Does Batch Normalization Help Binary Training?0
Computation-Efficient Quantization Method for Deep Neural Networks0
Countering Adversarial Examples: Combining Input Transformation and Noisy Training0
Covariance Recovery for One-Bit Sampled Data With Time-Varying Sampling Thresholds-Part I: Stationary Signals0
Covering Numbers for Deep ReLU Networks with Applications to Function Approximation and Nonparametric Regression0
COVIDLite: A depth-wise separable deep neural network with white balance and CLAHE for detection of COVID-190
A Structurally Regularized Convolutional Neural Network for Image Classification using Wavelet-based SubBand Decomposition0
A Feature-map Discriminant Perspective for Pruning Deep Neural Networks0
CPTQuant -- A Novel Mixed Precision Post-Training Quantization Techniques for Large Language Models0
CPT-V: A Contrastive Approach to Post-Training Quantization of Vision Transformers0
Accurate Block Quantization in LLMs with Outliers0
CQ-VAE: Coordinate Quantized VAE for Uncertainty Estimation with Application to Disk Shape Analysis from Lumbar Spine MRI Images0
3U-EdgeAI: Ultra-Low Memory Training, Ultra-Low BitwidthQuantization, and Ultra-Low Latency Acceleration0
CRB Analysis for Mixed-ADC Based DOA Estimation0
Computational Complexity Evaluation of Neural Network Applications in Signal Processing0
Computability of Classification and Deep Learning: From Theoretical Limits to Practical Feasibility through Quantization0
Crop Disease Classification using Support Vector Machines with Green Chromatic Coordinate (GCC) and Attention based feature extraction for IoT based Smart Agricultural Applications0
Cross-Dataset Propensity Estimation for Debiasing Recommender Systems0
Cross-Layer Discrete Concept Discovery for Interpreting Language Models0
Cross-Layer Optimization for Fault-Tolerant Deep Learning0
Cross-Modal Discrete Representation Learning0
Atleus: Accelerating Transformers on the Edge Enabled by 3D Heterogeneous Manycore Architectures0
A Structurally Regularized CNN Architecture via Adaptive Subband Decomposition0
Compress, Then Prompt: Improving Accuracy-Efficiency Trade-off of LLM Inference with Transferable Prompt0
Compress Polyphone Pronunciation Prediction Model with Shared Labels0
Show:102550
← PrevPage 41 of 197Next →

Benchmark Results

#ModelMetricClaimedVerifiedStatus
1FQ-ViT (ViT-L)Top-1 Accuracy (%)85.03Unverified
2FQ-ViT (ViT-B)Top-1 Accuracy (%)83.31Unverified
3FQ-ViT (Swin-B)Top-1 Accuracy (%)82.97Unverified
4FQ-ViT (Swin-S)Top-1 Accuracy (%)82.71Unverified
5FQ-ViT (DeiT-B)Top-1 Accuracy (%)81.2Unverified
6FQ-ViT (Swin-T)Top-1 Accuracy (%)80.51Unverified
7FQ-ViT (DeiT-S)Top-1 Accuracy (%)79.17Unverified
8Xception W8A8Top-1 Accuracy (%)78.97Unverified
9ADLIK-MO-ResNet50-W4A4Top-1 Accuracy (%)77.88Unverified
10ADLIK-MO-ResNet50-W3A4Top-1 Accuracy (%)77.34Unverified
#ModelMetricClaimedVerifiedStatus
13DCNN_VIVA_3MAP160,327.04Unverified
2DTQMAP0.79Unverified
#ModelMetricClaimedVerifiedStatus
1OutEffHop-Bert_basePerplexity6.3Unverified
2OutEffHop-Bert_basePerplexity6.21Unverified
#ModelMetricClaimedVerifiedStatus
1Accuracy98.13Unverified
#ModelMetricClaimedVerifiedStatus
1Accuracy92.92Unverified
#ModelMetricClaimedVerifiedStatus
1SSD ResNet50 V1 FPN 640x640MAP34.3Unverified
#ModelMetricClaimedVerifiedStatus
1TAR @ FAR=1e-495.13Unverified
#ModelMetricClaimedVerifiedStatus
1TAR @ FAR=1e-496.38Unverified
#ModelMetricClaimedVerifiedStatus
13DCNN_VIVA_5All84,809,664Unverified
#ModelMetricClaimedVerifiedStatus
1Accuracy99.8Unverified