SOTAVerified

Quantization

Quantization is a promising technique to reduce the computation cost of neural network training, which can replace high-cost floating-point numbers (e.g., float32) with low-cost fixed-point numbers (e.g., int8/int16).

Source: Adaptive Precision Training: Quantify Back Propagation in Neural Networks with Fixed-point Numbers

Papers

Showing 39013925 of 4925 papers

TitleStatusHype
Bit Error Robustness for Energy-Efficient DNN AcceleratorsCode0
Lattice Representation Learning0
Multi-Class Uncertainty Calibration via Mutual Information Maximization-based BinningCode0
IDF++: Analyzing and Improving Integer Discrete Flows for Lossless Compression0
Exploiting Weight Redundancy in CNNs: Beyond Pruning and Quantization0
Efficient Integer-Arithmetic-Only Convolutional Neural NetworksCode0
Unified Analysis of Stochastic Gradient Methods for Composite Convex and Smooth Optimization0
DEED: A General Quantization Scheme for Communication Efficiency in Bits0
COVIDLite: A depth-wise separable deep neural network with white balance and CLAHE for detection of COVID-190
Efficient Execution of Quantized Deep Learning Models: A Compiler Approach0
Federated Learning With Quantized Global Model Updates0
Some useful approximations for calculation of directivities of multibeam power patterns of large planar arrays0
Universally Quantized Neural Compression0
Quantization of Acoustic Model Parameters in Automatic Speech Recognition Framework0
CNN Acceleration by Low-rank Approximation with Quantized Factors0
Neural gradients are near-lognormal: improved quantized and sparse training0
HyperFlow: Representing 3D Objects as SurfacesCode0
Sparsity Turns Adversarial: Energy and Latency Attacks on Deep Neural Networks0
O(1) Communication for Distributed SGD through Two-Level Gradient Averaging0
Noisy One-bit Compressed Sensing with Side-Information0
Knowledge Distillation: A Survey0
Neural Network Activation Quantization with Bitwise Information BottlenecksCode0
Generative Design of Hardware-aware DNNs0
An Overview of Neural Network Compression0
Exploring the Potential of Low-bit Training of Convolutional Neural Networks0
Show:102550
← PrevPage 157 of 197Next →

Benchmark Results

#ModelMetricClaimedVerifiedStatus
1FQ-ViT (ViT-L)Top-1 Accuracy (%)85.03Unverified
2FQ-ViT (ViT-B)Top-1 Accuracy (%)83.31Unverified
3FQ-ViT (Swin-B)Top-1 Accuracy (%)82.97Unverified
4FQ-ViT (Swin-S)Top-1 Accuracy (%)82.71Unverified
5FQ-ViT (DeiT-B)Top-1 Accuracy (%)81.2Unverified
6FQ-ViT (Swin-T)Top-1 Accuracy (%)80.51Unverified
7FQ-ViT (DeiT-S)Top-1 Accuracy (%)79.17Unverified
8Xception W8A8Top-1 Accuracy (%)78.97Unverified
9ADLIK-MO-ResNet50-W4A4Top-1 Accuracy (%)77.88Unverified
10ADLIK-MO-ResNet50-W3A4Top-1 Accuracy (%)77.34Unverified
#ModelMetricClaimedVerifiedStatus
13DCNN_VIVA_3MAP160,327.04Unverified
2DTQMAP0.79Unverified
#ModelMetricClaimedVerifiedStatus
1OutEffHop-Bert_basePerplexity6.3Unverified
2OutEffHop-Bert_basePerplexity6.21Unverified
#ModelMetricClaimedVerifiedStatus
1Accuracy98.13Unverified
#ModelMetricClaimedVerifiedStatus
1Accuracy92.92Unverified
#ModelMetricClaimedVerifiedStatus
1SSD ResNet50 V1 FPN 640x640MAP34.3Unverified
#ModelMetricClaimedVerifiedStatus
1TAR @ FAR=1e-495.13Unverified
#ModelMetricClaimedVerifiedStatus
1TAR @ FAR=1e-496.38Unverified
#ModelMetricClaimedVerifiedStatus
13DCNN_VIVA_5All84,809,664Unverified
#ModelMetricClaimedVerifiedStatus
1Accuracy99.8Unverified