SOTAVerified

Quantization

Quantization is a promising technique to reduce the computation cost of neural network training, which can replace high-cost floating-point numbers (e.g., float32) with low-cost fixed-point numbers (e.g., int8/int16).

Source: Adaptive Precision Training: Quantify Back Propagation in Neural Networks with Fixed-point Numbers

Papers

Showing 36513700 of 4925 papers

TitleStatusHype
OL-DN: Online learning based dual-domain network for HEVC intra frame quality enhancement0
On Accelerating Edge AI: Optimizing Resource-Constrained Environments0
On Adaptive Transmission for Distributed Detection in Energy Harvesting Wireless Sensor Networks with Limited Fusion Center Feedback0
On a Relation Between the Rate-Distortion Function and Optimal Transport0
On Calibration of Modern Quantized Efficient Neural Networks0
Once Quantized for All: Progressively Searching for Quantized Compact Models0
On-demand Quantization for Green Federated Generative Diffusion in Mobile Edge Networks0
On Designing Modulation for Over-the-Air Computation -- Part I: Noise-Aware Design0
On-device AI: Quantization-aware Training of Transformers in Time-Series0
On-Device Learning with Binary Neural Networks0
On-Device Qwen2.5: Efficient LLM Inference with Model Compression and Hardware Acceleration0
On Distributed Quantization for Classification0
One-Bit-Aided Modulo Sampling for DOA Estimation0
One-Bit Channel Estimation for IRS-aided Millimeter-Wave Massive MU-MISO System0
One-Bit Direct Position Determination of Narrowband Gaussian Signals0
One-bit LFMCW Radar: Spectrum Analysis and Target Detection0
One-Bit MIMO Detection: From Global Maximum-Likelihood Detector to Amplitude Retrieval Approach0
One-Bit OFDM Receivers via Deep Learning0
One-Bit Quantization and Sparsification for Multiclass Linear Classification with Strong Regularization0
One-Bit Sigma-Delta DFRC Waveform Design: Using Quantization Noise for Radar Probing0
One-Bit Sigma-Delta modulation on the circle0
One-Bit Target Detection in Collocated MIMO Radar and Performance Degradation Analysis0
On Effects of Compression with Hyperdimensional Computing in Distributed Randomized Neural Networks0
On Efficient Constructions of Checkpoints0
One-Index Vector Quantization Based Adversarial Attack on Image Classification0
One Model for All Quantization: A Quantized Network Supporting Hot-Swap Bit-Width Adjustment0
One-pass Multiple Conformer and Foundation Speech Systems Compression and Quantization Using An All-in-one Neural Model0
One QuantLLM for ALL: Fine-tuning Quantized LLMs Once for Efficient Deployments0
One-Shot Model for Mixed-Precision Quantization0
One Weight Bitwidth to Rule Them All0
On Hardening DNNs against Noisy Computations0
On Image Registration and Subpixel Estimation0
On Large-Scale Retrieval: Binary or n-ary Coding?0
On Leaky-Integrate-and Fire as Spike-Train-Quantization Operator on Dirac-Superimposed Continuous-Time Signals0
Online Categorical Subspace Learning for Sketching Big Data with Misses0
Online Embedding Compression for Text Classification using Low Rank Matrix Factorization0
Online Learned Continual Compression with Stacked Quantization Modules0
Online Meta Adaptation for Variable-Rate Learned Image Compression0
Online Model Compression for Federated Learning with Large Models0
Online Product Quantization0
Online Signature Recognition: A Biologically Inspired Feature Vector Splitting Approach0
On Minimizing Symbol Error Rate Over Fading Channels with Low-Resolution Quantization0
On Neural Architecture Search for Resource-Constrained Hardware Platforms0
On Noncommutative Quantum Mechanics and the Black-Scholes Model0
On One-Bit Quantization0
On the Arrow of Inference0
On Periodic Functions as Regularizers for Quantization of Neural Networks0
On Quantizing Implicit Neural Representations0
On Randomized Distributed Coordinate Descent with Quantized Updates0
On Recursive State Estimation for Linear State-Space Models Having Quantized Output Data0
Show:102550
← PrevPage 74 of 99Next →

Benchmark Results

#ModelMetricClaimedVerifiedStatus
1FQ-ViT (ViT-L)Top-1 Accuracy (%)85.03Unverified
2FQ-ViT (ViT-B)Top-1 Accuracy (%)83.31Unverified
3FQ-ViT (Swin-B)Top-1 Accuracy (%)82.97Unverified
4FQ-ViT (Swin-S)Top-1 Accuracy (%)82.71Unverified
5FQ-ViT (DeiT-B)Top-1 Accuracy (%)81.2Unverified
6FQ-ViT (Swin-T)Top-1 Accuracy (%)80.51Unverified
7FQ-ViT (DeiT-S)Top-1 Accuracy (%)79.17Unverified
8Xception W8A8Top-1 Accuracy (%)78.97Unverified
9ADLIK-MO-ResNet50-W4A4Top-1 Accuracy (%)77.88Unverified
10ADLIK-MO-ResNet50-W3A4Top-1 Accuracy (%)77.34Unverified
#ModelMetricClaimedVerifiedStatus
13DCNN_VIVA_3MAP160,327.04Unverified
2DTQMAP0.79Unverified
#ModelMetricClaimedVerifiedStatus
1OutEffHop-Bert_basePerplexity6.3Unverified
2OutEffHop-Bert_basePerplexity6.21Unverified
#ModelMetricClaimedVerifiedStatus
1Accuracy98.13Unverified
#ModelMetricClaimedVerifiedStatus
1Accuracy92.92Unverified
#ModelMetricClaimedVerifiedStatus
1SSD ResNet50 V1 FPN 640x640MAP34.3Unverified
#ModelMetricClaimedVerifiedStatus
1TAR @ FAR=1e-495.13Unverified
#ModelMetricClaimedVerifiedStatus
1TAR @ FAR=1e-496.38Unverified
#ModelMetricClaimedVerifiedStatus
13DCNN_VIVA_5All84,809,664Unverified
#ModelMetricClaimedVerifiedStatus
1Accuracy99.8Unverified