SOTAVerified

Quantization

Quantization is a promising technique to reduce the computation cost of neural network training, which can replace high-cost floating-point numbers (e.g., float32) with low-cost fixed-point numbers (e.g., int8/int16).

Source: Adaptive Precision Training: Quantify Back Propagation in Neural Networks with Fixed-point Numbers

Papers

Showing 19512000 of 4925 papers

TitleStatusHype
FAT: An In-Memory Accelerator with Fast Addition for Ternary Weight Neural Networks0
Fast top-K Cosine Similarity Search through XOR-Friendly Binary Quantization on GPUs0
Compact and Robust Deep Learning Architecture for Fluorescence Lifetime Imaging and FPGA Implementation0
Generative Design of Hardware-aware DNNs0
Generative Diffusion Models for Lattice Field Theory0
Fast Template Evaluation with Vector Quantization0
Generative QoE Modeling: A Lightweight Approach for Telecom Networks0
Generative Semantic Communication for Text-to-Speech Synthesis0
A Reconstruction-Computation-Quantization (RCQ) Approach to Node Operations in LDPC Decoding0
A Deep Hashing Learning Network0
FastSGD: A Fast Compressed SGD Framework for Distributed Machine Learning0
FastQuery: Communication-efficient Embedding Table Query for Private LLM Inference0
A reconfigurable neural network ASIC for detector front-end data compression at the HL-LHC0
Geometry and clustering with metrics derived from separable Bregman divergences0
Fast Orthogonal Projection Based on Kronecker Product0
Getting Free Bits Back from Rotational Symmetries in LLMs0
Fast on-line signature recognition based on VQ with time modeling0
GHN-Q: Parameter Prediction for Unseen Quantized Convolutional Architectures via Graph Hypernetworks0
A Reconfigurable Dual-Mode Tracking SAR ADC without Analog Subtraction0
Acceleration of Convolutional Neural Network Using FFT-Based Split Convolutions0
3D Pathfinding and Collision Avoidance Using Uneven Search-space Quantization and Visual Cone Search0
Givens Coordinate Descent Methods for Rotation Matrix Learning in Trainable Embedding Indexes0
Synaptic Modulation using Interspike Intervals Increases Energy Efficiency of Spiking Neural Networks0
FastMamba: A High-Speed and Efficient Mamba Accelerator on FPGA with Accurate Quantization0
Communication-efficient Variance-reduced Stochastic Gradient Descent0
Global synchronization of multi-agent systems with nonlinear interactions0
Goal-oriented compression for L_p-norm-type goal functions: Application to power consumption scheduling0
Goal-Oriented Quantization: Analysis, Design, and Application to Resource Allocation0
Fast Low-rank Representation based Spatial Pyramid Matching for Image Classification0
GOBO: Quantizing Attention-Based NLP Models for Low Latency and Energy Efficient Inference0
Communication-Efficient Split Learning via Adaptive Feature-Wise Compression0
Going Further With Winograd Convolutions: Tap-Wise Quantization for Efficient Inference on 4x4 Tile0
Google USM: Scaling Automatic Speech Recognition Beyond 100 Languages0
Fast learning rates with heavy-tailed losses0
gpcgc: a green point cloud geometry coding method0
GPLQ: A General, Practical, and Lightning QAT Method for Vision Transformers0
Fast Large-Scale Discrete Optimization Based on Principal Coordinate Descent0
GPTQT: Quantize Large Language Models Twice to Push the Efficiency0
Fast Jet Tagging with MLP-Mixers on FPGAs0
Fast Inference of Tree Ensembles on ARM Devices0
Communication Efficient SGD via Gradient Sampling With Bayes Prior0
GQ-Net: Training Quantization-Friendly Deep Networks0
GQSA: Group Quantization and Sparsity for Accelerating Large Language Model Inference0
GradFreeBits: Gradient Free Bit Allocation for Dynamic Low Precision Neural Networks0
AddNet: Deep Neural Networks Using FPGA-Optimized Multipliers0
WaveQ: Gradient-Based Deep Quantization of Neural Networks through Sinusoidal Adaptive Regularization0
Gradient Based Method for the Fusion of Lattice Quantizers0
Gradient-Based Post-Training Quantization: Challenging the Status Quo0
Gradient Descent Quantizes ReLU Network Features0
Fast Implementation of 4-bit Convolutional Neural Networks for Mobile Devices0
Show:102550
← PrevPage 40 of 99Next →

Benchmark Results

#ModelMetricClaimedVerifiedStatus
1FQ-ViT (ViT-L)Top-1 Accuracy (%)85.03Unverified
2FQ-ViT (ViT-B)Top-1 Accuracy (%)83.31Unverified
3FQ-ViT (Swin-B)Top-1 Accuracy (%)82.97Unverified
4FQ-ViT (Swin-S)Top-1 Accuracy (%)82.71Unverified
5FQ-ViT (DeiT-B)Top-1 Accuracy (%)81.2Unverified
6FQ-ViT (Swin-T)Top-1 Accuracy (%)80.51Unverified
7FQ-ViT (DeiT-S)Top-1 Accuracy (%)79.17Unverified
8Xception W8A8Top-1 Accuracy (%)78.97Unverified
9ADLIK-MO-ResNet50-W4A4Top-1 Accuracy (%)77.88Unverified
10ADLIK-MO-ResNet50-W3A4Top-1 Accuracy (%)77.34Unverified
#ModelMetricClaimedVerifiedStatus
13DCNN_VIVA_3MAP160,327.04Unverified
2DTQMAP0.79Unverified
#ModelMetricClaimedVerifiedStatus
1OutEffHop-Bert_basePerplexity6.3Unverified
2OutEffHop-Bert_basePerplexity6.21Unverified
#ModelMetricClaimedVerifiedStatus
1Accuracy98.13Unverified
#ModelMetricClaimedVerifiedStatus
1Accuracy92.92Unverified
#ModelMetricClaimedVerifiedStatus
1SSD ResNet50 V1 FPN 640x640MAP34.3Unverified
#ModelMetricClaimedVerifiedStatus
1TAR @ FAR=1e-495.13Unverified
#ModelMetricClaimedVerifiedStatus
1TAR @ FAR=1e-496.38Unverified
#ModelMetricClaimedVerifiedStatus
13DCNN_VIVA_5All84,809,664Unverified
#ModelMetricClaimedVerifiedStatus
1Accuracy99.8Unverified