SOTAVerified

Quantization

Quantization is a promising technique to reduce the computation cost of neural network training, which can replace high-cost floating-point numbers (e.g., float32) with low-cost fixed-point numbers (e.g., int8/int16).

Source: Adaptive Precision Training: Quantify Back Propagation in Neural Networks with Fixed-point Numbers

Papers

Showing 21262150 of 4925 papers

TitleStatusHype
Collaborative Automotive Radar Sensing via Mixed-Precision Distributed Array Completion0
Vector Quantization for Deep-Learning-Based CSI Feedback in Massive MIMO Systems0
Approaching Rate-Distortion Limits in Neural Compression with Lattice Transform Coding0
FlowVQTalker: High-Quality Emotional Talking Face Generation through Normalizing Flow and Quantization0
What Makes Quantization for Large Language Models Hard? An Empirical Study from the Lens of Perturbation0
QuantTune: Optimizing Model Quantization with Adaptive Outlier-Driven Fine Tuning0
Enhancing Multimodal Unified Representations for Cross Modal Generalization0
Micro-Fracture Detection in Photovoltaic Cells with Hardware-Constrained Devices and Computer Vision0
The Impact of Quantization on the Robustness of Transformer-based Text Classifiers0
Algorithm-Hardware Co-Design of Distribution-Aware Logarithmic-Posit Encodings for Efficient DNN InferenceCode0
LoCoDL: Communication-Efficient Distributed Learning with Local Training and Compression0
On-demand Quantization for Green Federated Generative Diffusion in Mobile Edge Networks0
Adaptive Integrate-and-Fire Time Encoding Machine with Quantization0
EasyQuant: An Efficient Data-free Quantization Algorithm for LLMs0
Design of Stochastic Quantizers for Privacy Preservation0
VQSynery: Robust Drug Synergy Prediction With Vector Quantization Mechanism0
Deep-Learned Compression for Radio-Frequency Signal Classification0
FlowPrecision: Advancing FPGA-Based Real-Time Fluid Flow Estimation with Linear Quantization0
Neural Network Assisted Lifting Steps For Improved Fully Scalable Lossy Image Compression in JPEG 2000Code0
Towards efficient deep autoencoders for multivariate time series anomaly detection0
Better Schedules for Low Precision Training of Deep Neural Networks0
A Hierarchical Federated Learning Approach for the Internet of Things0
On the Compressibility of Quantized Large Language Models0
Extracting Usable Predictions from Quantized Networks through Uncertainty Quantification for OOD DetectionCode0
BasedAI: A decentralized P2P network for Zero Knowledge Large Language Models (ZK-LLMs)0
Show:102550
← PrevPage 86 of 197Next →

Benchmark Results

#ModelMetricClaimedVerifiedStatus
1FQ-ViT (ViT-L)Top-1 Accuracy (%)85.03Unverified
2FQ-ViT (ViT-B)Top-1 Accuracy (%)83.31Unverified
3FQ-ViT (Swin-B)Top-1 Accuracy (%)82.97Unverified
4FQ-ViT (Swin-S)Top-1 Accuracy (%)82.71Unverified
5FQ-ViT (DeiT-B)Top-1 Accuracy (%)81.2Unverified
6FQ-ViT (Swin-T)Top-1 Accuracy (%)80.51Unverified
7FQ-ViT (DeiT-S)Top-1 Accuracy (%)79.17Unverified
8Xception W8A8Top-1 Accuracy (%)78.97Unverified
9ADLIK-MO-ResNet50-W4A4Top-1 Accuracy (%)77.88Unverified
10ADLIK-MO-ResNet50-W3A4Top-1 Accuracy (%)77.34Unverified
#ModelMetricClaimedVerifiedStatus
13DCNN_VIVA_3MAP160,327.04Unverified
2DTQMAP0.79Unverified
#ModelMetricClaimedVerifiedStatus
1OutEffHop-Bert_basePerplexity6.3Unverified
2OutEffHop-Bert_basePerplexity6.21Unverified
#ModelMetricClaimedVerifiedStatus
1Accuracy98.13Unverified
#ModelMetricClaimedVerifiedStatus
1Accuracy92.92Unverified
#ModelMetricClaimedVerifiedStatus
1SSD ResNet50 V1 FPN 640x640MAP34.3Unverified
#ModelMetricClaimedVerifiedStatus
1TAR @ FAR=1e-495.13Unverified
#ModelMetricClaimedVerifiedStatus
1TAR @ FAR=1e-496.38Unverified
#ModelMetricClaimedVerifiedStatus
13DCNN_VIVA_5All84,809,664Unverified
#ModelMetricClaimedVerifiedStatus
1Accuracy99.8Unverified