SOTAVerified

Quantization

Quantization is a promising technique to reduce the computation cost of neural network training, which can replace high-cost floating-point numbers (e.g., float32) with low-cost fixed-point numbers (e.g., int8/int16).

Source: Adaptive Precision Training: Quantify Back Propagation in Neural Networks with Fixed-point Numbers

Papers

Showing 13511375 of 4925 papers

TitleStatusHype
A Novel Structure-Agnostic Multi-Objective Approach for Weight-Sharing Compression in Deep Neural Networks0
Can General-Purpose Large Language Models Generalize to English-Thai Machine Translation ?0
A Novel Physics-based Channel Model for Reconfigurable Intelligent Surface-assisted Multi-user Communication Systems0
Adaptive Joint Optimization for 3D Reconstruction with Differentiable Rendering0
Elastic Significant Bit Quantization and Acceleration for Deep Neural Networks0
Cancer Subtyping via Embedded Unsupervised Learning on Transcriptomics Data0
CAMBI: Contrast-aware Multiscale Banding Index0
A Novel Light Field Coding Scheme Based on Deep Belief Network & Weighted Binary Images for Additive Layered Displays0
CALM: Co-evolution of Algorithms and Language Model for Automatic Heuristic Design0
A Novel Hybrid Precoder With Low-Resolution Phase Shifters and Fronthaul Capacity Limitation0
Adaptive Integrate-and-Fire Time Encoding Machine with Quantization0
Cactus Mechanisms: Optimal Differential Privacy Mechanisms in the Large-Composition Regime0
A Novel Framework for Image-to-image Translation and Image Compression0
CacheQuant: Comprehensively Accelerated Diffusion Models0
A Novel Chaotic Uniform Quantizer for Speech Coding0
Accelerated Distance Computation with Encoding Tree for High Dimensional Data0
ELMGS: Enhancing memory and computation scaLability through coMpression for 3D Gaussian Splatting0
Emotion Recognition Using Speaker Cues0
A Novel Audio Representation for Music Genre Identification in MIR0
CA3D: Convolutional-Attentional 3D Nets for Efficient Video Activity Recognition on the Edge0
Byzantine-Resilient Secure Federated Learning0
A Novel Approach to Quantized Matrix Completion Using Huber Loss Measure0
Adaptive Dither Voting for Robust Spatial Verification0
Bullion: A Column Store for Machine Learning0
Building an Efficiency Pipeline: Commutativity and Cumulativeness of Efficiency Operators for Transformers0
Show:102550
← PrevPage 55 of 197Next →

Benchmark Results

#ModelMetricClaimedVerifiedStatus
1FQ-ViT (ViT-L)Top-1 Accuracy (%)85.03Unverified
2FQ-ViT (ViT-B)Top-1 Accuracy (%)83.31Unverified
3FQ-ViT (Swin-B)Top-1 Accuracy (%)82.97Unverified
4FQ-ViT (Swin-S)Top-1 Accuracy (%)82.71Unverified
5FQ-ViT (DeiT-B)Top-1 Accuracy (%)81.2Unverified
6FQ-ViT (Swin-T)Top-1 Accuracy (%)80.51Unverified
7FQ-ViT (DeiT-S)Top-1 Accuracy (%)79.17Unverified
8Xception W8A8Top-1 Accuracy (%)78.97Unverified
9ADLIK-MO-ResNet50-W4A4Top-1 Accuracy (%)77.88Unverified
10ADLIK-MO-ResNet50-W3A4Top-1 Accuracy (%)77.34Unverified
#ModelMetricClaimedVerifiedStatus
13DCNN_VIVA_3MAP160,327.04Unverified
2DTQMAP0.79Unverified
#ModelMetricClaimedVerifiedStatus
1OutEffHop-Bert_basePerplexity6.3Unverified
2OutEffHop-Bert_basePerplexity6.21Unverified
#ModelMetricClaimedVerifiedStatus
1Accuracy98.13Unverified
#ModelMetricClaimedVerifiedStatus
1Accuracy92.92Unverified
#ModelMetricClaimedVerifiedStatus
1SSD ResNet50 V1 FPN 640x640MAP34.3Unverified
#ModelMetricClaimedVerifiedStatus
1TAR @ FAR=1e-495.13Unverified
#ModelMetricClaimedVerifiedStatus
1TAR @ FAR=1e-496.38Unverified
#ModelMetricClaimedVerifiedStatus
13DCNN_VIVA_5All84,809,664Unverified
#ModelMetricClaimedVerifiedStatus
1Accuracy99.8Unverified