SOTAVerified

Quantization

Quantization is a promising technique to reduce the computation cost of neural network training, which can replace high-cost floating-point numbers (e.g., float32) with low-cost fixed-point numbers (e.g., int8/int16).

Source: Adaptive Precision Training: Quantify Back Propagation in Neural Networks with Fixed-point Numbers

Papers

Showing 34763500 of 4925 papers

TitleStatusHype
Adaptive Dither Voting for Robust Spatial Verification0
Adaptive Integrate-and-Fire Time Encoding Machine with Quantization0
Adaptive Joint Optimization for 3D Reconstruction with Differentiable Rendering0
Adaptive Low-Precision Training for Embeddings in Click-Through Rate Prediction0
Adaptive Periodic Averaging: A Practical Approach to Reducing Communication in Distributed Learning0
Adaptive Precision Training: Quantify Back Propagation in Neural Networks with Fixed-point Numbers0
Adaptive Proximal Gradient Methods for Structured Neural Networks0
Adaptive Quantization for Deep Neural Network0
Adaptive Quantization for Key Generation in Low-Power Wide-Area Networks0
Adaptive Quantization of Model Updates for Communication-Efficient Federated Learning0
Adaptive Quantization of Neural Networks0
Adaptive Quantization Resolution and Power Control for Federated Learning over Cell-free Networks0
Adaptive quantization with mixed-precision based on low-cost proxy0
Adaptive Resolution Inference (ARI): Energy-Efficient Machine Learning for Internet of Things0
Adaptive Resource Allocation for Semantic Communication Networks0
Adaptive Sample-space & Adaptive Probability coding: a neural-network based approach for compression0
Adaptive Training of Random Mapping for Data Quantization0
Adaptive Transmission for Distributed Detection in Energy Harvesting Wireless Sensor Networks0
Adaptive Wireless Image Semantic Transmission and Over-The-Air Testing0
Adaptive Wireless Image Semantic Transmission: Design, Simulation, and Prototype Validation0
AdaptivFloat: A Floating-point based Data Type for Resilient Deep Learning Inference0
AdaQAT: Adaptive Bit-Width Quantization-Aware Training0
A Data and Compute Efficient Design for Limited-Resources Deep Learning0
AdderNet and its Minimalist Hardware Design for Energy-Efficient Artificial Intelligence0
Additive Quantization for Extreme Vector Compression0
Show:102550
← PrevPage 140 of 197Next →

Benchmark Results

#ModelMetricClaimedVerifiedStatus
1FQ-ViT (ViT-L)Top-1 Accuracy (%)85.03Unverified
2FQ-ViT (ViT-B)Top-1 Accuracy (%)83.31Unverified
3FQ-ViT (Swin-B)Top-1 Accuracy (%)82.97Unverified
4FQ-ViT (Swin-S)Top-1 Accuracy (%)82.71Unverified
5FQ-ViT (DeiT-B)Top-1 Accuracy (%)81.2Unverified
6FQ-ViT (Swin-T)Top-1 Accuracy (%)80.51Unverified
7FQ-ViT (DeiT-S)Top-1 Accuracy (%)79.17Unverified
8Xception W8A8Top-1 Accuracy (%)78.97Unverified
9ADLIK-MO-ResNet50-W4A4Top-1 Accuracy (%)77.88Unverified
10ADLIK-MO-ResNet50-W3A4Top-1 Accuracy (%)77.34Unverified
#ModelMetricClaimedVerifiedStatus
13DCNN_VIVA_3MAP160,327.04Unverified
2DTQMAP0.79Unverified
#ModelMetricClaimedVerifiedStatus
1OutEffHop-Bert_basePerplexity6.3Unverified
2OutEffHop-Bert_basePerplexity6.21Unverified
#ModelMetricClaimedVerifiedStatus
1Accuracy98.13Unverified
#ModelMetricClaimedVerifiedStatus
1Accuracy92.92Unverified
#ModelMetricClaimedVerifiedStatus
1SSD ResNet50 V1 FPN 640x640MAP34.3Unverified
#ModelMetricClaimedVerifiedStatus
1TAR @ FAR=1e-495.13Unverified
#ModelMetricClaimedVerifiedStatus
1TAR @ FAR=1e-496.38Unverified
#ModelMetricClaimedVerifiedStatus
13DCNN_VIVA_5All84,809,664Unverified
#ModelMetricClaimedVerifiedStatus
1Accuracy99.8Unverified