SOTAVerified

Quantization

Quantization is a promising technique to reduce the computation cost of neural network training, which can replace high-cost floating-point numbers (e.g., float32) with low-cost fixed-point numbers (e.g., int8/int16).

Source: Adaptive Precision Training: Quantify Back Propagation in Neural Networks with Fixed-point Numbers

Papers

Showing 46514700 of 4925 papers

TitleStatusHype
The Neural Network Pushdown Automaton: Model, Stack and Learning SimulationsCode0
Weightless: Lossy Weight Encoding For Deep Neural Network CompressionCode0
ADaPTION: Toolbox and Benchmark for Training Convolutional Neural Networks with Reduced Numerical Precision Weights and Activation0
Quantized Memory-Augmented Neural Networks0
Unbounded cache model for online language modeling with open vocabularyCode0
Distribution-Preserving k-Anonymity0
Compressing Word Embeddings via Deep Compositional Code LearningCode0
Attacking Binarized Neural Networks0
SUT System Description for Anti-Spoofing 2017 Challenge0
Efficient Inferencing of Compressed Deep Neural Networks0
Towards Effective Low-bitwidth Convolutional Neural NetworksCode0
Deep Hashing with Triplet Quantization Loss0
Quantization goes Polynomial0
Deep Learning as a Mixed Convex-Combinatorial Optimization ProblemCode0
Using the quantization error from Self-Organized Map (SOM) output for detecting critical variability in large bodies of image time series in less than a minute0
High Five: Improving Gesture Recognition by Embracing Uncertainty0
End-to-End Optimized Speech Coding with Deep Neural Networks0
A Survey of Model Compression and Acceleration for Deep Neural Networks0
Low Precision RNNs: Quantizing RNNs Without Losing Accuracy0
Vector Quantization using the Improved Differential Evolution Algorithm for Image Compression0
TensorQuant - A Simulation Toolbox for Deep Neural Network QuantizationCode0
STDP Based Pruning of Connections and Weight Quantization in Spiking Neural Networks for Energy Efficient Recognition0
Quantized Minimum Error Entropy Criterion0
Compressive Quantization for Fast Object Instance Search in Videos0
3D Surface Detail Enhancement From a Single Normal Map0
AnnArbor: Approximate Nearest Neighbors Using Arborescence Coding0
The model of an anomaly detector for HiLumi LHC magnets based on Recurrent Neural Networks and adaptive quantizationCode0
Compressing Low Precision Deep Neural Networks Using Sparsity-Induced Regularization in Ternary Networks0
E^2BoWs: An End-to-End Bag-of-Words Model via Deep Convolutional Neural Network0
Simulated Annealing for JPEG Quantization0
Towards Compact and Fast Neural Machine Translation Using a Combined Method0
Joint Maximum Purity Forest with Application to Image Super-ResolutionCode0
Performance Guaranteed Network Acceleration via High-Order Residual Quantization0
Neural Networks Compression for Language Modeling0
Deep Neural Network Capacity0
SUBIC: A supervised, structured binary code for image search0
Learning Accurate Low-Bit Deep Neural Networks with Stochastic QuantizationCode0
Monocular Depth Estimation with Hierarchical Fusion of Dilated CNNs and Soft-Weighted-Sum InferenceCode0
ZipML: Training Linear Models with End-to-End Low Precision, and a Little Bit of Deep Learning0
Learning Bag-of-Features Pooling for Deep Convolutional Neural NetworksCode0
Extremely Low Bit Neural Network: Squeeze the Last Bit Out with ADMM0
Model compression as constrained optimization, with application to neural nets. Part II: quantization0
A multi-layer image representation using Regularized Residual Quantization: application to compression and denoising0
Model compression as constrained optimization, with application to neural nets. Part I: general framework0
Learning Deep Binary Descriptor With Multi-Quantization0
Deep Visual-Semantic Quantization for Efficient Image Retrieval0
Weighted-Entropy-Based Quantization for Deep Neural Networks0
Product Split Trees0
Balanced Quantization: An Effective and Efficient Approach to Quantized Neural Networks0
Representation Learning using Event-based STDP0
Show:102550
← PrevPage 94 of 99Next →

Benchmark Results

#ModelMetricClaimedVerifiedStatus
1FQ-ViT (ViT-L)Top-1 Accuracy (%)85.03Unverified
2FQ-ViT (ViT-B)Top-1 Accuracy (%)83.31Unverified
3FQ-ViT (Swin-B)Top-1 Accuracy (%)82.97Unverified
4FQ-ViT (Swin-S)Top-1 Accuracy (%)82.71Unverified
5FQ-ViT (DeiT-B)Top-1 Accuracy (%)81.2Unverified
6FQ-ViT (Swin-T)Top-1 Accuracy (%)80.51Unverified
7FQ-ViT (DeiT-S)Top-1 Accuracy (%)79.17Unverified
8Xception W8A8Top-1 Accuracy (%)78.97Unverified
9ADLIK-MO-ResNet50-W4A4Top-1 Accuracy (%)77.88Unverified
10ADLIK-MO-ResNet50-W3A4Top-1 Accuracy (%)77.34Unverified
#ModelMetricClaimedVerifiedStatus
13DCNN_VIVA_3MAP160,327.04Unverified
2DTQMAP0.79Unverified
#ModelMetricClaimedVerifiedStatus
1OutEffHop-Bert_basePerplexity6.3Unverified
2OutEffHop-Bert_basePerplexity6.21Unverified
#ModelMetricClaimedVerifiedStatus
1Accuracy98.13Unverified
#ModelMetricClaimedVerifiedStatus
1Accuracy92.92Unverified
#ModelMetricClaimedVerifiedStatus
1SSD ResNet50 V1 FPN 640x640MAP34.3Unverified
#ModelMetricClaimedVerifiedStatus
1TAR @ FAR=1e-495.13Unverified
#ModelMetricClaimedVerifiedStatus
1TAR @ FAR=1e-496.38Unverified
#ModelMetricClaimedVerifiedStatus
13DCNN_VIVA_5All84,809,664Unverified
#ModelMetricClaimedVerifiedStatus
1Accuracy99.8Unverified