SOTAVerified

Quantization

Quantization is a promising technique to reduce the computation cost of neural network training, which can replace high-cost floating-point numbers (e.g., float32) with low-cost fixed-point numbers (e.g., int8/int16).

Source: Adaptive Precision Training: Quantify Back Propagation in Neural Networks with Fixed-point Numbers

Papers

Showing 37763800 of 4925 papers

TitleStatusHype
Automatic Gain Control Design for Dynamic Visible Light Communication Systems0
Geometric Total Variation for Image Vectorization, Zooming and Pixel Art DepixelizingCode1
Communication-Efficient Federated Learning via Optimal Client Sampling0
Flexible framework for audio reconstructionCode0
On the Impact of Lossy Image and Video Compression on the Performance of Deep Convolutional Neural Network Architectures0
WrapNet: Neural Net Inference with Ultra-Low-Resolution Arithmetic0
End-to-end Learning of Compressible Features0
Analysis and Optimization for RIS-Aided Multi-Pair Communications Relying on Statistical CSI0
A Post-coder Feedback Approach to Overcome Training Asymmetry in MIMO-TDD0
Byzantine-Resilient Secure Federated Learning0
Differentiable Joint Pruning and Quantization for Hardware Efficiency0
HMQ: Hardware Friendly Mixed Precision Quantization Block for CNNsCode1
Search What You Want: Barrier Panelty NAS for Mixed Precision Quantization0
Exploiting vulnerabilities of deep neural networks for privacy protectionCode0
Resolution Switchable Networks for Runtime Efficient Image RecognitionCode1
DBQ: A Differentiable Branch Quantizer for Lightweight Deep Neural Networks0
FSpiNN: An Optimization Framework for Memory- and Energy-Efficient Spiking Neural Networks0
Training with reduced precision of a support vector machine model for text classification0
eSampling: Energy Harvesting ADCs0
Device-Robust Acoustic Scene Classification Based on Two-Stage Categorization and Data AugmentationCode1
Channel-Level Variable Quantization Network for Deep Image CompressionCode1
Compression strategies and space-conscious representations for deep neural networks0
A General Family of Stochastic Proximal Gradient Methods for Deep Learning0
Image De-Quantization Using Generative Models as Priors0
Finding Non-Uniform Quantization Schemes using Multi-Task Gaussian ProcessesCode0
Show:102550
← PrevPage 152 of 197Next →

Benchmark Results

#ModelMetricClaimedVerifiedStatus
1FQ-ViT (ViT-L)Top-1 Accuracy (%)85.03Unverified
2FQ-ViT (ViT-B)Top-1 Accuracy (%)83.31Unverified
3FQ-ViT (Swin-B)Top-1 Accuracy (%)82.97Unverified
4FQ-ViT (Swin-S)Top-1 Accuracy (%)82.71Unverified
5FQ-ViT (DeiT-B)Top-1 Accuracy (%)81.2Unverified
6FQ-ViT (Swin-T)Top-1 Accuracy (%)80.51Unverified
7FQ-ViT (DeiT-S)Top-1 Accuracy (%)79.17Unverified
8Xception W8A8Top-1 Accuracy (%)78.97Unverified
9ADLIK-MO-ResNet50-W4A4Top-1 Accuracy (%)77.88Unverified
10ADLIK-MO-ResNet50-W3A4Top-1 Accuracy (%)77.34Unverified
#ModelMetricClaimedVerifiedStatus
13DCNN_VIVA_3MAP160,327.04Unverified
2DTQMAP0.79Unverified
#ModelMetricClaimedVerifiedStatus
1OutEffHop-Bert_basePerplexity6.3Unverified
2OutEffHop-Bert_basePerplexity6.21Unverified
#ModelMetricClaimedVerifiedStatus
1Accuracy98.13Unverified
#ModelMetricClaimedVerifiedStatus
1Accuracy92.92Unverified
#ModelMetricClaimedVerifiedStatus
1SSD ResNet50 V1 FPN 640x640MAP34.3Unverified
#ModelMetricClaimedVerifiedStatus
1TAR @ FAR=1e-495.13Unverified
#ModelMetricClaimedVerifiedStatus
1TAR @ FAR=1e-496.38Unverified
#ModelMetricClaimedVerifiedStatus
13DCNN_VIVA_5All84,809,664Unverified
#ModelMetricClaimedVerifiedStatus
1Accuracy99.8Unverified