SOTAVerified

Quantization

Quantization is a promising technique to reduce the computation cost of neural network training, which can replace high-cost floating-point numbers (e.g., float32) with low-cost fixed-point numbers (e.g., int8/int16).

Source: Adaptive Precision Training: Quantify Back Propagation in Neural Networks with Fixed-point Numbers

Papers

Showing 15511575 of 4925 papers

TitleStatusHype
End-to-End Optimized Speech Coding with Deep Neural Networks0
End-to-end Quantized Training via Log-Barrier Extensions0
End-to-End Rate-Distortion Optimization for Bi-Directional Learned Video Compression0
An Implementation of Vector Quantization using the Genetic Algorithm Approach0
EDAS: Efficient and Differentiable Architecture Search0
Adaptive Asymmetric Label-guided Hashing for Multimedia Search0
Energy-Aware LLMs: A step towards sustainable AI for downstream applications0
Energy awareness in low precision neural networks0
Energy Efficiency Maximization Precoding for Quantized Massive MIMO Systems0
Energy Efficiency Optimization for Millimeter Wave System with Resolution-Adaptive ADCs0
Energy-Efficient Distributed Learning Algorithms for Coarsely Quantized Signals0
Energy-efficient Knowledge Distillation for Spiking Neural Networks0
Energy Efficient Learning with Low Resolution Stochastic Domain Wall Synapse Based Deep Neural Networks0
Energy-Efficient Respiratory Anomaly Detection in Premature Newborn Infants0
Energy-Efficient Transformer Inference: Optimization Strategies for Time Series Classification0
Engineering the Neural Automatic Passenger Counter0
ECG Biometric Authentication Using Self-Supervised Learning for IoT Edge Sensors0
Enhanced Bayesian Compression via Deep Reinforcement Learning0
BlueLM-V-3B: Algorithm and System Co-Design for Multimodal Large Language Models on Mobile Devices0
Single-path Bit Sharing for Automatic Loss-aware Model Compression0
Fast binary embeddings, and quantized compressed sensing with structured matrices0
Enhancement Of Coded Speech Using a Mask-Based Post-Filter0
Fast, Compact, and High Quality LSTM-RNN Based Statistical Parametric Speech Synthesizers for Mobile Devices0
Block-Wise Dynamic-Precision Neural Network Training Acceleration via Online Quantization Sensitivity Analytics0
Reclaiming Residual Knowledge: A Novel Paradigm to Low-Bit Quantization0
Show:102550
← PrevPage 63 of 197Next →

Benchmark Results

#ModelMetricClaimedVerifiedStatus
1FQ-ViT (ViT-L)Top-1 Accuracy (%)85.03Unverified
2FQ-ViT (ViT-B)Top-1 Accuracy (%)83.31Unverified
3FQ-ViT (Swin-B)Top-1 Accuracy (%)82.97Unverified
4FQ-ViT (Swin-S)Top-1 Accuracy (%)82.71Unverified
5FQ-ViT (DeiT-B)Top-1 Accuracy (%)81.2Unverified
6FQ-ViT (Swin-T)Top-1 Accuracy (%)80.51Unverified
7FQ-ViT (DeiT-S)Top-1 Accuracy (%)79.17Unverified
8Xception W8A8Top-1 Accuracy (%)78.97Unverified
9ADLIK-MO-ResNet50-W4A4Top-1 Accuracy (%)77.88Unverified
10ADLIK-MO-ResNet50-W3A4Top-1 Accuracy (%)77.34Unverified
#ModelMetricClaimedVerifiedStatus
13DCNN_VIVA_3MAP160,327.04Unverified
2DTQMAP0.79Unverified
#ModelMetricClaimedVerifiedStatus
1OutEffHop-Bert_basePerplexity6.3Unverified
2OutEffHop-Bert_basePerplexity6.21Unverified
#ModelMetricClaimedVerifiedStatus
1Accuracy98.13Unverified
#ModelMetricClaimedVerifiedStatus
1Accuracy92.92Unverified
#ModelMetricClaimedVerifiedStatus
1SSD ResNet50 V1 FPN 640x640MAP34.3Unverified
#ModelMetricClaimedVerifiedStatus
1TAR @ FAR=1e-495.13Unverified
#ModelMetricClaimedVerifiedStatus
1TAR @ FAR=1e-496.38Unverified
#ModelMetricClaimedVerifiedStatus
13DCNN_VIVA_5All84,809,664Unverified
#ModelMetricClaimedVerifiedStatus
1Accuracy99.8Unverified