SOTAVerified

Quantization

Quantization is a promising technique to reduce the computation cost of neural network training, which can replace high-cost floating-point numbers (e.g., float32) with low-cost fixed-point numbers (e.g., int8/int16).

Source: Adaptive Precision Training: Quantify Back Propagation in Neural Networks with Fixed-point Numbers

Papers

Showing 21262150 of 4925 papers

TitleStatusHype
Expectation maximization transfer learning and its application for bionic hand prostheses0
Expand-and-Quantize: Unsupervised Semantic Segmentation Using High-Dimensional Space and Product Quantization0
Codec-ASR: Training Performant Automatic Speech Recognition Systems with Discrete Speech Representations0
A Precision-Scalable RISC-V DNN Processor with On-Device Learning Capability at the Extreme Edge0
Adaptive Wireless Image Semantic Transmission: Design, Simulation, and Prototype Validation0
eXmY: A Data Type and Technique for Arbitrary Bit Precision Quantization0
CodeBrain: Impute Any Brain MRI via Instance-specific Scalar-quantized Codes0
Examining the Role and Limits of Batchnorm Optimization to Mitigate Diverse Hardware-noise in In-memory Computing0
A Practical Mixed Precision Algorithm for Post-Training Quantization0
Exact Recovery of Sparse Binary Vectors from Generalized Linear Measurements0
Codebook based Audio Feature Representation for Music Information Retrieval0
Exact Bias Correction and Covariance Estimation for Stereo Vision0
Codage \'echelonnable \`a granularit\'e fine de la parole : Application au codeur G.729 (Fine granularity scalable speech coding: Application to the G.729 coder) [in French]0
Adaptive Wireless Image Semantic Transmission and Over-The-Air Testing0
Event-Triggered Quantized Average Consensus via Mass Summation0
Distributed Inference with Sparse and Quantized Communication0
Event Retrieval in Large Video Collections with Circulant Temporal Encoding0
Cocktail: Chunk-Adaptive Mixed-Precision Quantization for Long-Context LLM Inference0
Approximation speed of quantized vs. unquantized ReLU neural networks and beyond0
Eventor: An Efficient Event-Based Monocular Multi-View Stereo Accelerator on FPGA Platform0
Event-Based Bispectral Photometry Using Temporally Modulated Illumination0
Evaluation of quality measures for color quantization0
Evaluation of Linear Implicit Quantized State System method for analyzing mission performance of power systems0
CoAst: Validation-Free Contribution Assessment for Federated Learning based on Cross-Round Valuation0
Approximation of functions with one-bit neural networks0
Show:102550
← PrevPage 86 of 197Next →

Benchmark Results

#ModelMetricClaimedVerifiedStatus
1FQ-ViT (ViT-L)Top-1 Accuracy (%)85.03Unverified
2FQ-ViT (ViT-B)Top-1 Accuracy (%)83.31Unverified
3FQ-ViT (Swin-B)Top-1 Accuracy (%)82.97Unverified
4FQ-ViT (Swin-S)Top-1 Accuracy (%)82.71Unverified
5FQ-ViT (DeiT-B)Top-1 Accuracy (%)81.2Unverified
6FQ-ViT (Swin-T)Top-1 Accuracy (%)80.51Unverified
7FQ-ViT (DeiT-S)Top-1 Accuracy (%)79.17Unverified
8Xception W8A8Top-1 Accuracy (%)78.97Unverified
9ADLIK-MO-ResNet50-W4A4Top-1 Accuracy (%)77.88Unverified
10ADLIK-MO-ResNet50-W3A4Top-1 Accuracy (%)77.34Unverified
#ModelMetricClaimedVerifiedStatus
13DCNN_VIVA_3MAP160,327.04Unverified
2DTQMAP0.79Unverified
#ModelMetricClaimedVerifiedStatus
1OutEffHop-Bert_basePerplexity6.3Unverified
2OutEffHop-Bert_basePerplexity6.21Unverified
#ModelMetricClaimedVerifiedStatus
1Accuracy98.13Unverified
#ModelMetricClaimedVerifiedStatus
1Accuracy92.92Unverified
#ModelMetricClaimedVerifiedStatus
1SSD ResNet50 V1 FPN 640x640MAP34.3Unverified
#ModelMetricClaimedVerifiedStatus
1TAR @ FAR=1e-495.13Unverified
#ModelMetricClaimedVerifiedStatus
1TAR @ FAR=1e-496.38Unverified
#ModelMetricClaimedVerifiedStatus
13DCNN_VIVA_5All84,809,664Unverified
#ModelMetricClaimedVerifiedStatus
1Accuracy99.8Unverified