SOTAVerified

Quantization

Quantization is a promising technique to reduce the computation cost of neural network training, which can replace high-cost floating-point numbers (e.g., float32) with low-cost fixed-point numbers (e.g., int8/int16).

Source: Adaptive Precision Training: Quantify Back Propagation in Neural Networks with Fixed-point Numbers

Papers

Showing 40514100 of 4925 papers

TitleStatusHype
Error Correcting Output Codes Improve Probability Estimation and Adversarial Robustness of Deep Neural NetworksCode0
Generalization Error Analysis of Quantized Compressive Learning0
Random Projections with Asymmetric Quantization0
The Synthesis of XNOR Recurrent Neural Networks with Stochastic Logic0
MetaQuant: Learning to Quantize by Learning to Penetrate Non-differentiable QuantizationCode1
Coresets for Archetypal AnalysisCode0
Normalization Helps Training of Quantized LSTMCode0
A binary-activation, multi-level weight RNN and training algorithm for ADC-/DAC-free and noise-resilient processing-in-memory inference with eNVM0
Semi-Relaxed Quantization with DropBits: Training Low-Bit Neural Networks via Bit-wise Regularization0
QKD: Quantization-aware Knowledge Distillation0
Neural Network-Inspired Analog-to-Digital Conversion to Achieve Super-Resolution with Low-Precision RRAM Devices0
Two-Stage Learning for Uplink Channel Estimation in One-Bit Massive MIMO0
Model-Aware Deep Architectures for One-Bit Compressive Variational AutoencodingCode0
Music Source Separation in the Waveform Domain0
A SOT-MRAM-based Processing-In-Memory Engine for Highly Compressed DNN Implementation0
Pyramid Vector Quantization and Bit Level Sparsity in Weights for Efficient Neural Networks Inference0
Quantization NetworksCode0
Learning Hierarchical Discrete Linguistic Units from Visually-Grounded Speech0
AddNet: Deep Neural Networks Using FPGA-Optimized Multipliers0
On the Discrepancy between the Theoretical Analysis and Practical Implementations of Compressed Communication for Distributed Deep LearningCode0
IFQ-Net: Integrated Fixed-point Quantization Networks for Embedded Vision0
Online Learned Continual Compression with Adaptive Quantization ModulesCode1
CoopNet: Cooperative Convolutional Neural Network for Low-Power MCUsCode0
vqSGD: Vector Quantized Stochastic Gradient Descent0
Efficient Hardware Implementation of Incremental Learning and Inference on Chip0
Loss Aware Post-training QuantizationCode0
Transductive Zero-Shot Hashing for Multilabel Image RetrievalCode0
Data Efficient Stagewise Knowledge DistillationCode0
One-Bit Sigma-Delta modulation on the circle0
The Canonical Distortion Measure for Vector Quantization and Function Approximation0
Quantization-based Bermudan option pricing in the FX world0
What Do Compressed Deep Neural Networks Forget?Code0
DupNet: Towards Very Tiny Quantized CNN with Improved Accuracy for Face Detection0
Scientific Image Restoration AnywhereCode1
Hyper-Sphere Quantization: Communication-Efficient SGD for Federated LearningCode0
Norm-Explicit Quantization: Improving Vector Quantization for Maximum Inner Product SearchCode0
Iteratively Training Look-Up Tables for Network Quantization0
HAWQ-V2: Hessian Aware trace-Weighted Quantization of Neural NetworksCode1
Effectiveness of self-supervised pre-training for speech recognitionCode1
ConveRT: Efficient and Accurate Conversational Representations from TransformersCode1
Multiple yield curve modelling with CBI processes0
A Programmable Approach to Neural Network CompressionCode0
Post-Training 4-bit Quantization on Embedding Tables0
Ternary MobileNets via Per-Layer Hybrid Filter Banks0
Comprehensive SNN Compression Using ADMM Optimization and Activity RegularizationCode0
Memory Requirement Reduction of Deep Neural Networks Using Low-bit Quantization of Parameters0
On Distributed Quantization for Classification0
Adaptive Precision Training: Quantify Back Propagation in Neural Networks with Fixed-point Numbers0
LFZip: Lossy compression of multivariate floating-point time series data via improved predictionCode0
MulCode: A Multiplicative Multi-way Model for Compressing Neural Language Model0
Show:102550
← PrevPage 82 of 99Next →

Benchmark Results

#ModelMetricClaimedVerifiedStatus
1FQ-ViT (ViT-L)Top-1 Accuracy (%)85.03Unverified
2FQ-ViT (ViT-B)Top-1 Accuracy (%)83.31Unverified
3FQ-ViT (Swin-B)Top-1 Accuracy (%)82.97Unverified
4FQ-ViT (Swin-S)Top-1 Accuracy (%)82.71Unverified
5FQ-ViT (DeiT-B)Top-1 Accuracy (%)81.2Unverified
6FQ-ViT (Swin-T)Top-1 Accuracy (%)80.51Unverified
7FQ-ViT (DeiT-S)Top-1 Accuracy (%)79.17Unverified
8Xception W8A8Top-1 Accuracy (%)78.97Unverified
9ADLIK-MO-ResNet50-W4A4Top-1 Accuracy (%)77.88Unverified
10ADLIK-MO-ResNet50-W3A4Top-1 Accuracy (%)77.34Unverified
#ModelMetricClaimedVerifiedStatus
13DCNN_VIVA_3MAP160,327.04Unverified
2DTQMAP0.79Unverified
#ModelMetricClaimedVerifiedStatus
1OutEffHop-Bert_basePerplexity6.3Unverified
2OutEffHop-Bert_basePerplexity6.21Unverified
#ModelMetricClaimedVerifiedStatus
1Accuracy98.13Unverified
#ModelMetricClaimedVerifiedStatus
1Accuracy92.92Unverified
#ModelMetricClaimedVerifiedStatus
1SSD ResNet50 V1 FPN 640x640MAP34.3Unverified
#ModelMetricClaimedVerifiedStatus
1TAR @ FAR=1e-495.13Unverified
#ModelMetricClaimedVerifiedStatus
1TAR @ FAR=1e-496.38Unverified
#ModelMetricClaimedVerifiedStatus
13DCNN_VIVA_5All84,809,664Unverified
#ModelMetricClaimedVerifiedStatus
1Accuracy99.8Unverified