SOTAVerified

Quantization

Quantization is a promising technique to reduce the computation cost of neural network training, which can replace high-cost floating-point numbers (e.g., float32) with low-cost fixed-point numbers (e.g., int8/int16).

Source: Adaptive Precision Training: Quantify Back Propagation in Neural Networks with Fixed-point Numbers

Papers

Showing 17261750 of 4925 papers

TitleStatusHype
Dynamic Predictive Sampling Analog to Digital Converter for Sparse Signal Sensing0
Blended Coarse Gradient Descent for Full Quantization of Deep Neural Networks0
ADaPTION: Toolbox and Benchmark for Training Convolutional Neural Networks with Reduced Numerical Precision Weights and Activation0
FQ-Conv: Fully Quantized Convolution for Efficient and Accurate Inference0
Adaptive Error-Bounded Hierarchical Matrices for Efficient Neural Network Compression0
Bi-ViT: Pushing the Limit of Vision Transformer Quantization0
FastSGD: A Fast Compressed SGD Framework for Distributed Machine Learning0
Dynamic Cell Modeling of Li-Ion Polymer Batteries for Precise SOC Estimation in Power-Needy Autonomous Electric Vehicles0
DyBit: Dynamic Bit-Precision Numbers for Efficient Quantized Neural Network Inference0
An Experimental Study of Reduced-Voltage Operation in Modern FPGAs for Neural Network Acceleration0
DV-Det: Efficient 3D Point Cloud Object Detection with Dynamic Voxelization0
Fast top-K Cosine Similarity Search through XOR-Friendly Binary Quantization on GPUs0
FAT: An In-Memory Accelerator with Fast Addition for Ternary Weight Neural Networks0
DupNet: Towards Very Tiny Quantized CNN with Improved Accuracy for Face Detection0
FATNN: Fast and Accurate Ternary Neural Networks0
Fault-Tolerant Four-Dimensional Constellation for Coherent Optical Transmission Systems0
BitTTS: Highly Compact Text-to-Speech Using 1.58-bit Quantization and Weight Indexing0
FBI: Fingerprinting models with Benign Inputs0
Compact recurrent neural networks for acoustic event detection on low-energy low-complexity platforms0
FBQuant: FeedBack Quantization for Large Language Models0
FCN-Pose: A Pruned and Quantized CNN for Robot Pose Estimation for Constrained Devices0
An Experimental Study: Assessing the Combined Framework of WavLM and BEST-RQ for Text-to-Speech Synthesis0
FD Cell-Free mMIMO: Analysis and Optimization0
Dual Precision Quantization for Efficient and Accurate Deep Neural Networks Inference0
An Exact Quantized Decentralized Gradient Descent Algorithm0
Show:102550
← PrevPage 70 of 197Next →

Benchmark Results

#ModelMetricClaimedVerifiedStatus
1FQ-ViT (ViT-L)Top-1 Accuracy (%)85.03Unverified
2FQ-ViT (ViT-B)Top-1 Accuracy (%)83.31Unverified
3FQ-ViT (Swin-B)Top-1 Accuracy (%)82.97Unverified
4FQ-ViT (Swin-S)Top-1 Accuracy (%)82.71Unverified
5FQ-ViT (DeiT-B)Top-1 Accuracy (%)81.2Unverified
6FQ-ViT (Swin-T)Top-1 Accuracy (%)80.51Unverified
7FQ-ViT (DeiT-S)Top-1 Accuracy (%)79.17Unverified
8Xception W8A8Top-1 Accuracy (%)78.97Unverified
9ADLIK-MO-ResNet50-W4A4Top-1 Accuracy (%)77.88Unverified
10ADLIK-MO-ResNet50-W3A4Top-1 Accuracy (%)77.34Unverified
#ModelMetricClaimedVerifiedStatus
13DCNN_VIVA_3MAP160,327.04Unverified
2DTQMAP0.79Unverified
#ModelMetricClaimedVerifiedStatus
1OutEffHop-Bert_basePerplexity6.3Unverified
2OutEffHop-Bert_basePerplexity6.21Unverified
#ModelMetricClaimedVerifiedStatus
1Accuracy98.13Unverified
#ModelMetricClaimedVerifiedStatus
1Accuracy92.92Unverified
#ModelMetricClaimedVerifiedStatus
1SSD ResNet50 V1 FPN 640x640MAP34.3Unverified
#ModelMetricClaimedVerifiedStatus
1TAR @ FAR=1e-495.13Unverified
#ModelMetricClaimedVerifiedStatus
1TAR @ FAR=1e-496.38Unverified
#ModelMetricClaimedVerifiedStatus
13DCNN_VIVA_5All84,809,664Unverified
#ModelMetricClaimedVerifiedStatus
1Accuracy99.8Unverified