SOTAVerified

Quantization

Quantization is a promising technique to reduce the computation cost of neural network training, which can replace high-cost floating-point numbers (e.g., float32) with low-cost fixed-point numbers (e.g., int8/int16).

Source: Adaptive Precision Training: Quantify Back Propagation in Neural Networks with Fixed-point Numbers

Papers

Showing 28012850 of 4925 papers

TitleStatusHype
FlatENN: Train Flat for Enhanced Fault Tolerance of Quantized Deep Neural Networks0
Publishing Efficient On-device Models Increases Adversarial Vulnerability0
Hyperspherical Loss-Aware Ternary Quantization0
Development of a Thermodynamics of Human Cognition and Human Culture0
Hyperspherical Quantization: Toward Smaller and More Accurate Models0
EuclidNets: An Alternative Operation for Efficient Inference of Deep Learning Models0
Cross-Dataset Propensity Estimation for Debiasing Recommender Systems0
Training Integer-Only Deep Recurrent Neural Networks0
Automatic Network Adaptation for Ultra-Low Uniform-Precision Quantization0
Towards Neural Variational Monte Carlo That Scales Linearly with System Size0
CSMPQ:Class Separability Based Mixed-Precision Quantization0
Redistribution of Weights and Activations for AdderNet Quantization0
Model Predictive Control for Neuromimetic Quantized Systems0
Atrous Space Bender U-Net (ASBU-Net/LogiNet)0
Joint SPX-VIX calibration with Gaussian polynomial volatility models: deep pricing with quantization hints0
NAWQ-SR: A Hybrid-Precision NPU Engine for Efficient On-Device Super-Resolution0
Huber-energy measure quantization0
Towards Hardware-Specific Automatic Compression of Neural Networks0
Image Compression with Product Quantized Masked Image Modeling0
Efficient Speech Representation Learning with Low-Bit Quantization0
Adaptive Low-Precision Training for Embeddings in Click-Through Rate Prediction0
ResFed: Communication Efficient Federated Learning by Transmitting Deep Compressed Residuals0
Error-aware Quantization through Noise Tempering0
QVIP: An ILP-based Formal Verification Approach for Quantized Neural NetworksCode0
Vertical Layering of Quantized Neural Networks for Heterogeneous Inference0
Genie: Show Me the Data for QuantizationCode0
TinyKG: Memory-Efficient Training Framework for Knowledge Graph Neural Recommender Systems0
QEBVerif: Quantization Error Bound Verification of Neural NetworksCode0
CSQ: Growing Mixed-Precision Quantization Scheme with Bi-level Continuous Sparsification0
Quantized Wasserstein Procrustes Alignment of Word Embedding Spaces0
QFT: Post-training quantization via fast joint finetuning of all degrees of freedom0
Make RepVGG Greater Again: A Quantization-aware ApproachCode0
Device Interoperability for Learned Image Compression with Weights and Activations Quantization0
Shining light on data: Geometric data analysis through quantum dynamics0
Robust Task-Specific Beamforming with Low-Resolution ADCs for Power-Efficient Hybrid MIMO Receivers0
Quadapter: Adapter for GPT-2 Quantization0
Quantization-aware Interval Bound Propagation for Training Certifiably Robust Quantized Neural NetworksCode0
Distributed Computation of Exact Average Degree and Network Size in Finite Number of Steps under Quantized Communication0
Quantized control of non-Lipschitz nonlinear systems: a novel control framework with prescribed transient performance and lower design complexity0
Entropy-Driven Mixed-Precision Quantization for Deep Network Design0
Class-based Quantization for Neural Networks0
Homology-constrained vector quantization entropy regularizerCode0
Vector Quantization Methods for Access Point Placement in Cell-Free Massive MIMO Systems0
A 65nm 8b-Activation 8b-Weight SRAM-Based Charge-Domain Computing-in-Memory Macro Using A Fully-Parallel Analog Adder Network and A Single-ADC Interface0
RIS-Assisted Self-Interference Mitigation for In-Band Full-Duplex Transceivers0
High-Perceptual Quality JPEG Decoding via Posterior Sampling0
Learning Low-Rank Representations for Model Compression0
Understanding and Improving Knowledge Distillation for Quantization-Aware Training of Large Transformer EncodersCode0
A Closed-loop Sleep Modulation System with FPGA-Accelerated Deep Learning0
Deep Learning-Aided Perturbation Model-Based Fiber Nonlinearity Compensation0
Show:102550
← PrevPage 57 of 99Next →

Benchmark Results

#ModelMetricClaimedVerifiedStatus
1FQ-ViT (ViT-L)Top-1 Accuracy (%)85.03Unverified
2FQ-ViT (ViT-B)Top-1 Accuracy (%)83.31Unverified
3FQ-ViT (Swin-B)Top-1 Accuracy (%)82.97Unverified
4FQ-ViT (Swin-S)Top-1 Accuracy (%)82.71Unverified
5FQ-ViT (DeiT-B)Top-1 Accuracy (%)81.2Unverified
6FQ-ViT (Swin-T)Top-1 Accuracy (%)80.51Unverified
7FQ-ViT (DeiT-S)Top-1 Accuracy (%)79.17Unverified
8Xception W8A8Top-1 Accuracy (%)78.97Unverified
9ADLIK-MO-ResNet50-W4A4Top-1 Accuracy (%)77.88Unverified
10ADLIK-MO-ResNet50-W3A4Top-1 Accuracy (%)77.34Unverified
#ModelMetricClaimedVerifiedStatus
13DCNN_VIVA_3MAP160,327.04Unverified
2DTQMAP0.79Unverified
#ModelMetricClaimedVerifiedStatus
1OutEffHop-Bert_basePerplexity6.3Unverified
2OutEffHop-Bert_basePerplexity6.21Unverified
#ModelMetricClaimedVerifiedStatus
1Accuracy98.13Unverified
#ModelMetricClaimedVerifiedStatus
1Accuracy92.92Unverified
#ModelMetricClaimedVerifiedStatus
1SSD ResNet50 V1 FPN 640x640MAP34.3Unverified
#ModelMetricClaimedVerifiedStatus
1TAR @ FAR=1e-495.13Unverified
#ModelMetricClaimedVerifiedStatus
1TAR @ FAR=1e-496.38Unverified
#ModelMetricClaimedVerifiedStatus
13DCNN_VIVA_5All84,809,664Unverified
#ModelMetricClaimedVerifiedStatus
1Accuracy99.8Unverified