SOTAVerified

Quantization

Quantization is a promising technique to reduce the computation cost of neural network training, which can replace high-cost floating-point numbers (e.g., float32) with low-cost fixed-point numbers (e.g., int8/int16).

Source: Adaptive Precision Training: Quantify Back Propagation in Neural Networks with Fixed-point Numbers

Papers

Showing 32513300 of 4925 papers

TitleStatusHype
Improving Robustness Against Stealthy Weight Bit-Flip Attacks by Output Code MatchingCode0
SceneSqueezer: Learning To Compress Scene for Camera Relocalization0
Instance-Aware Dynamic Neural Network QuantizationCode0
Mutual Quantization for Cross-Modal Search With Noisy Labels0
Quantization-Aware Deep Optics for Diffractive Snapshot Hyperspectral Imaging0
Data-Free Network Compression via Parametric Non-Uniform Mixed Precision Quantization0
Mr.BiQ: Post-Training Non-Uniform Quantization Based on Minimizing the Reconstruction Error0
Croesus: Multi-Stage Processing and Transactions for Video-Analytics in Edge-Cloud Systems0
Automatic Mixed-Precision Quantization Search of BERT0
Studying the Interplay between Information Loss and Operation Loss in Representations for Classification0
End-to-End Autoencoder Communications with Optimized Interference Suppression0
Speedup deep learning models on GPU by taking advantage of efficient unstructured pruning and bit-width reduction0
HiKonv: High Throughput Quantized Convolution With Novel Bit-wise Management and Computation0
BMPQ: Bit-Gradient Sensitivity Driven Mixed-Precision Quantization of DNNs from Scratch0
Stochastic Learning Equation using Monotone Increasing Resolution of Quantization0
Distilling the Knowledge of Romanian BERTs Using Multiple TeachersCode0
Manifold learning via quantum dynamics0
Accurate Neural Training with 4-bit Matrix Multiplications at Standard Formats0
Nearest neighbor search with compact codes: A decoder perspective0
Approximation of functions with one-bit neural networks0
TAFA: Design Automation of Analog Mixed-Signal FIR Filters Using Time Approximation Architecture0
On Recursive State Estimation for Linear State-Space Models Having Quantized Output Data0
Boosted Dense Retriever0
Modeling Image Quantization Tradeoffs for Optimal Compression0
Illumination and Temperature-Aware Multispectral Networks for Edge-Computing-Enabled Pedestrian Detection0
Efficient Batch Homomorphic Encryption for Vertically Federated XGBoost0
FastSGD: A Fast Compressed SGD Framework for Distributed Machine Learning0
Implicit Neural Representations for Image Compression0
Neural Network Quantization for Efficient Inference: A Survey0
A Generalized Zero-Shot Quantization of Deep Convolutional Neural Networks via Learned Weights Statistics0
A comparison study of CNN denoisers on PRNU extraction0
Towards Low-loss 1-bit Quantization of User-item Representations for Top-K Recommendation0
Equal Bits: Enforcing Equally Distributed Binary Network WeightsCode0
High-Resolution WiFi Imaging with Reconfigurable Intelligent Surfaces0
Adaptive Proximal Gradient Methods for Structured Neural Networks0
Exploration into Translation-Equivariant Image QuantizationCode0
Attribute Artifacts Removal for Geometry-based Point Cloud Compression0
Hardware-friendly Deep Learning by Network Quantization and Binarization0
Communication-Efficient Federated Learning via Quantized Compressed Sensing0
A Highly Effective Low-Rank Compression of Deep Neural Networks with Modified Beam-Search and Modified Stable Rank0
Mixed Precision of Quantization of Transformer Language Models for Speech Recognition0
Mixed Precision Low-bit Quantization of Neural Network Language Models for Speech Recognition0
Low-bit Quantization of Recurrent Neural Network Language Models Using Alternating Direction Methods of Multipliers0
Enabling Fast Deep Learning on Tiny Energy-Harvesting IoT Devices0
An Optimization Framework for Federated Edge Learning0
QNNVerifier: A Tool for Verifying Neural Networks using SMT-Based Model Checking0
A Novel Framework for Image-to-image Translation and Image Compression0
Accelerating Deep Learning with Dynamic Data Pruning0
HERO: Hessian-Enhanced Robust Optimization for Unifying and Improving Generalization and Quantization PerformanceCode0
Full-Duplex Massive MIMO Cellular Networks with Low Resolution ADC/DAC0
Show:102550
← PrevPage 66 of 99Next →

Benchmark Results

#ModelMetricClaimedVerifiedStatus
1FQ-ViT (ViT-L)Top-1 Accuracy (%)85.03Unverified
2FQ-ViT (ViT-B)Top-1 Accuracy (%)83.31Unverified
3FQ-ViT (Swin-B)Top-1 Accuracy (%)82.97Unverified
4FQ-ViT (Swin-S)Top-1 Accuracy (%)82.71Unverified
5FQ-ViT (DeiT-B)Top-1 Accuracy (%)81.2Unverified
6FQ-ViT (Swin-T)Top-1 Accuracy (%)80.51Unverified
7FQ-ViT (DeiT-S)Top-1 Accuracy (%)79.17Unverified
8Xception W8A8Top-1 Accuracy (%)78.97Unverified
9ADLIK-MO-ResNet50-W4A4Top-1 Accuracy (%)77.88Unverified
10ADLIK-MO-ResNet50-W3A4Top-1 Accuracy (%)77.34Unverified
#ModelMetricClaimedVerifiedStatus
13DCNN_VIVA_3MAP160,327.04Unverified
2DTQMAP0.79Unverified
#ModelMetricClaimedVerifiedStatus
1OutEffHop-Bert_basePerplexity6.3Unverified
2OutEffHop-Bert_basePerplexity6.21Unverified
#ModelMetricClaimedVerifiedStatus
1Accuracy98.13Unverified
#ModelMetricClaimedVerifiedStatus
1Accuracy92.92Unverified
#ModelMetricClaimedVerifiedStatus
1SSD ResNet50 V1 FPN 640x640MAP34.3Unverified
#ModelMetricClaimedVerifiedStatus
1TAR @ FAR=1e-495.13Unverified
#ModelMetricClaimedVerifiedStatus
1TAR @ FAR=1e-496.38Unverified
#ModelMetricClaimedVerifiedStatus
13DCNN_VIVA_5All84,809,664Unverified
#ModelMetricClaimedVerifiedStatus
1Accuracy99.8Unverified