SOTAVerified

Quantization

Quantization is a promising technique to reduce the computation cost of neural network training, which can replace high-cost floating-point numbers (e.g., float32) with low-cost fixed-point numbers (e.g., int8/int16).

Source: Adaptive Precision Training: Quantify Back Propagation in Neural Networks with Fixed-point Numbers

Papers

Showing 38013825 of 4925 papers

TitleStatusHype
Patch-wise Mixed-Precision Quantization of Vision Transformer0
A rescaling-invariant Lipschitz bound based on path-metrics for modern ReLU network parameterizations0
Pathology Image Compression with Pre-trained Autoencoders0
PathVQ: Reforming Computational Pathology Foundation Model for Whole Slide Image Analysis via Vector Quantization0
PCDVQ: Enhancing Vector Quantization for Large Language Models via Polar Coordinate Decoupling0
PECAN: A Product-Quantized Content Addressable Memory Network0
Perceptual Video Quality Prediction Emphasizing Chroma Distortions0
Performance Analysis of IRS-Assisted Cell-Free Communication0
Performance Analysis of Massive MIMO Multi-Way Relay Networks with Low-Resolution ADCs0
Performance Guaranteed Network Acceleration via High-Order Residual Quantization0
Performance of Cell-Free MmWave Massive MIMO Systems with Fronthaul Compression and DAC Quantization0
Performance Optimization for Variable Bitwidth Federated Learning in Wireless Networks0
Persistence Codebooks for Topological Data Analysis0
Personalized Federated Learning for Cellular VR: Online Learning and Dynamic Caching0
Per-Tensor Fixed-Point Quantization of the Back-Propagation Algorithm0
Phoenix: A Low-Precision Floating-Point Quantization Oriented Architecture for Convolutional Neural Networks0
Physics-aware Differentiable Discrete Codesign for Diffractive Optical Neural Networks0
Physics of Language Models: Part 3.3, Knowledge Capacity Scaling Laws0
Pieces of Eight: 8-bit Neural Machine Translation0
PikeLPN: Mitigating Overlooked Inefficiencies of Low-Precision Neural Networks0
Information Entropy Guided Height-aware Histogram for Quantization-friendly Pillar Feature Encoder0
PillarHist: A Quantization-aware Pillar Feature Encoder based on Height-aware Histogram0
PIM-QAT: Neural Network Quantization for Processing-In-Memory (PIM) Systems0
Pinball Loss Minimization for One-bit Compressive Sensing: Convex Models and Algorithms0
Pioneering 4-Bit FP Quantization for Diffusion Models: Mixup-Sign Quantization and Timestep-Aware Fine-Tuning0
Show:102550
← PrevPage 153 of 197Next →

Benchmark Results

#ModelMetricClaimedVerifiedStatus
1FQ-ViT (ViT-L)Top-1 Accuracy (%)85.03Unverified
2FQ-ViT (ViT-B)Top-1 Accuracy (%)83.31Unverified
3FQ-ViT (Swin-B)Top-1 Accuracy (%)82.97Unverified
4FQ-ViT (Swin-S)Top-1 Accuracy (%)82.71Unverified
5FQ-ViT (DeiT-B)Top-1 Accuracy (%)81.2Unverified
6FQ-ViT (Swin-T)Top-1 Accuracy (%)80.51Unverified
7FQ-ViT (DeiT-S)Top-1 Accuracy (%)79.17Unverified
8Xception W8A8Top-1 Accuracy (%)78.97Unverified
9ADLIK-MO-ResNet50-W4A4Top-1 Accuracy (%)77.88Unverified
10ADLIK-MO-ResNet50-W3A4Top-1 Accuracy (%)77.34Unverified
#ModelMetricClaimedVerifiedStatus
13DCNN_VIVA_3MAP160,327.04Unverified
2DTQMAP0.79Unverified
#ModelMetricClaimedVerifiedStatus
1OutEffHop-Bert_basePerplexity6.3Unverified
2OutEffHop-Bert_basePerplexity6.21Unverified
#ModelMetricClaimedVerifiedStatus
1Accuracy98.13Unverified
#ModelMetricClaimedVerifiedStatus
1Accuracy92.92Unverified
#ModelMetricClaimedVerifiedStatus
1SSD ResNet50 V1 FPN 640x640MAP34.3Unverified
#ModelMetricClaimedVerifiedStatus
1TAR @ FAR=1e-495.13Unverified
#ModelMetricClaimedVerifiedStatus
1TAR @ FAR=1e-496.38Unverified
#ModelMetricClaimedVerifiedStatus
13DCNN_VIVA_5All84,809,664Unverified
#ModelMetricClaimedVerifiedStatus
1Accuracy99.8Unverified