SOTAVerified

Quantization

Quantization is a promising technique to reduce the computation cost of neural network training, which can replace high-cost floating-point numbers (e.g., float32) with low-cost fixed-point numbers (e.g., int8/int16).

Source: Adaptive Precision Training: Quantify Back Propagation in Neural Networks with Fixed-point Numbers

Papers

Showing 23512375 of 4925 papers

TitleStatusHype
A White Paper on Neural Network Quantization0
Demystifying Singular Defects in Large Language Models0
Joint Neural Architecture Search and Quantization0
Joint Optimization of Rate, Distortion, and Decoding Energy for HEVC Intraframe Coding0
AlphaTuning: Quantization-Aware Parameter-Efficient Adaptation of Large-Scale Pre-Trained Language Models0
Bang for the Buck: Vector Search on Cloud CPUs0
Learning-Based Latency-Constrained Fronthaul Compression Optimization in C-RAN0
Joint Quantization and Pruning Neural Networks Approach: A Case Study on FSO Receivers0
Joint Sequential Fronthaul Quantization and Hardware Complexity Reduction in Uplink Cell-Free Massive MIMO Networks0
Joint SPX-VIX calibration with Gaussian polynomial volatility models: deep pricing with quantization hints0
Intelligent Fault Diagnosis of Type and Severity in Low-Frequency, Low Bit-Depth Signals0
Joint Texture and Geometry Optimization for RGB-D Reconstruction0
Joshua 4.0: Packing, PRO, and Paraphrases0
Deployment of Deep Neural Networks for Object Detection on Edge AI Devices with Runtime Optimization0
JPEG-LM: LLMs as Image Generators with Canonical Codec Representations0
JPEG Quantized Coefficient Recovery via DCT Domain Spatial-Frequential Transformer0
Integrating PHY Security Into NDN-IoT Networks By Exploiting MEC: Authentication Efficiency, Robustness, and Accuracy Enhancement0
Deep neural networks are robust to weight binarization and other non-linear distortions0
Dequantization of a signal from two parallel quantized observations0
AWEQ: Post-Training Quantization with Activation-Weight Equalization for Large Language Models0
Deep neural networks algorithms for stochastic control problems on finite horizon: convergence analysis0
KDLSQ-BERT: A Quantized Bert Combining Knowledge Distillation with Learned Step Size Quantization0
Kernel k-Medoids as General Vector Quantization0
Kernel Quantization for Efficient Network Compression0
Learning a Virtual Codec Based on Deep Convolutional Neural Network to Compress Image0
Show:102550
← PrevPage 95 of 197Next →

Benchmark Results

#ModelMetricClaimedVerifiedStatus
1FQ-ViT (ViT-L)Top-1 Accuracy (%)85.03Unverified
2FQ-ViT (ViT-B)Top-1 Accuracy (%)83.31Unverified
3FQ-ViT (Swin-B)Top-1 Accuracy (%)82.97Unverified
4FQ-ViT (Swin-S)Top-1 Accuracy (%)82.71Unverified
5FQ-ViT (DeiT-B)Top-1 Accuracy (%)81.2Unverified
6FQ-ViT (Swin-T)Top-1 Accuracy (%)80.51Unverified
7FQ-ViT (DeiT-S)Top-1 Accuracy (%)79.17Unverified
8Xception W8A8Top-1 Accuracy (%)78.97Unverified
9ADLIK-MO-ResNet50-W4A4Top-1 Accuracy (%)77.88Unverified
10ADLIK-MO-ResNet50-W3A4Top-1 Accuracy (%)77.34Unverified
#ModelMetricClaimedVerifiedStatus
13DCNN_VIVA_3MAP160,327.04Unverified
2DTQMAP0.79Unverified
#ModelMetricClaimedVerifiedStatus
1OutEffHop-Bert_basePerplexity6.3Unverified
2OutEffHop-Bert_basePerplexity6.21Unverified
#ModelMetricClaimedVerifiedStatus
1Accuracy98.13Unverified
#ModelMetricClaimedVerifiedStatus
1Accuracy92.92Unverified
#ModelMetricClaimedVerifiedStatus
1SSD ResNet50 V1 FPN 640x640MAP34.3Unverified
#ModelMetricClaimedVerifiedStatus
1TAR @ FAR=1e-495.13Unverified
#ModelMetricClaimedVerifiedStatus
1TAR @ FAR=1e-496.38Unverified
#ModelMetricClaimedVerifiedStatus
13DCNN_VIVA_5All84,809,664Unverified
#ModelMetricClaimedVerifiedStatus
1Accuracy99.8Unverified