SOTAVerified

Quantization

Quantization is a promising technique to reduce the computation cost of neural network training, which can replace high-cost floating-point numbers (e.g., float32) with low-cost fixed-point numbers (e.g., int8/int16).

Source: Adaptive Precision Training: Quantify Back Propagation in Neural Networks with Fixed-point Numbers

Papers

Showing 46514675 of 4925 papers

TitleStatusHype
Compressing Word Embeddings via Deep Compositional Code LearningCode0
SUT System Description for Anti-Spoofing 2017 Challenge0
Efficient Inferencing of Compressed Deep Neural Networks0
Towards Effective Low-bitwidth Convolutional Neural NetworksCode0
Attacking Binarized Neural Networks0
Quantization goes Polynomial0
Deep Hashing with Triplet Quantization Loss0
Deep Learning as a Mixed Convex-Combinatorial Optimization ProblemCode0
Using the quantization error from Self-Organized Map (SOM) output for detecting critical variability in large bodies of image time series in less than a minute0
End-to-End Optimized Speech Coding with Deep Neural Networks0
High Five: Improving Gesture Recognition by Embracing Uncertainty0
A Survey of Model Compression and Acceleration for Deep Neural Networks0
Low Precision RNNs: Quantizing RNNs Without Losing Accuracy0
Vector Quantization using the Improved Differential Evolution Algorithm for Image Compression0
TensorQuant - A Simulation Toolbox for Deep Neural Network QuantizationCode0
STDP Based Pruning of Connections and Weight Quantization in Spiking Neural Networks for Energy Efficient Recognition0
Quantized Minimum Error Entropy Criterion0
3D Surface Detail Enhancement From a Single Normal Map0
Compressive Quantization for Fast Object Instance Search in Videos0
AnnArbor: Approximate Nearest Neighbors Using Arborescence Coding0
The model of an anomaly detector for HiLumi LHC magnets based on Recurrent Neural Networks and adaptive quantizationCode0
Compressing Low Precision Deep Neural Networks Using Sparsity-Induced Regularization in Ternary Networks0
E^2BoWs: An End-to-End Bag-of-Words Model via Deep Convolutional Neural Network0
Simulated Annealing for JPEG Quantization0
Towards Compact and Fast Neural Machine Translation Using a Combined Method0
Show:102550
← PrevPage 187 of 197Next →

Benchmark Results

#ModelMetricClaimedVerifiedStatus
1FQ-ViT (ViT-L)Top-1 Accuracy (%)85.03Unverified
2FQ-ViT (ViT-B)Top-1 Accuracy (%)83.31Unverified
3FQ-ViT (Swin-B)Top-1 Accuracy (%)82.97Unverified
4FQ-ViT (Swin-S)Top-1 Accuracy (%)82.71Unverified
5FQ-ViT (DeiT-B)Top-1 Accuracy (%)81.2Unverified
6FQ-ViT (Swin-T)Top-1 Accuracy (%)80.51Unverified
7FQ-ViT (DeiT-S)Top-1 Accuracy (%)79.17Unverified
8Xception W8A8Top-1 Accuracy (%)78.97Unverified
9ADLIK-MO-ResNet50-W4A4Top-1 Accuracy (%)77.88Unverified
10ADLIK-MO-ResNet50-W3A4Top-1 Accuracy (%)77.34Unverified
#ModelMetricClaimedVerifiedStatus
13DCNN_VIVA_3MAP160,327.04Unverified
2DTQMAP0.79Unverified
#ModelMetricClaimedVerifiedStatus
1OutEffHop-Bert_basePerplexity6.3Unverified
2OutEffHop-Bert_basePerplexity6.21Unverified
#ModelMetricClaimedVerifiedStatus
1Accuracy98.13Unverified
#ModelMetricClaimedVerifiedStatus
1Accuracy92.92Unverified
#ModelMetricClaimedVerifiedStatus
1SSD ResNet50 V1 FPN 640x640MAP34.3Unverified
#ModelMetricClaimedVerifiedStatus
1TAR @ FAR=1e-495.13Unverified
#ModelMetricClaimedVerifiedStatus
1TAR @ FAR=1e-496.38Unverified
#ModelMetricClaimedVerifiedStatus
13DCNN_VIVA_5All84,809,664Unverified
#ModelMetricClaimedVerifiedStatus
1Accuracy99.8Unverified