SOTAVerified

Quantization

Quantization is a promising technique to reduce the computation cost of neural network training, which can replace high-cost floating-point numbers (e.g., float32) with low-cost fixed-point numbers (e.g., int8/int16).

Source: Adaptive Precision Training: Quantify Back Propagation in Neural Networks with Fixed-point Numbers

Papers

Showing 31013125 of 4925 papers

TitleStatusHype
Towards Semantic Communications: Deep Learning-Based Image Semantic Coding0
Towards Superior Quantization Accuracy: A Layer-sensitive Approach0
Towards the Limit of Network Quantization0
Towards Unified INT8 Training for Convolutional Neural Network0
Towards Unsupervised Speech Recognition and Synthesis with Quantized Speech Representation Learning0
Towards Variable and Coordinated Holistic Co-Speech Motion Generation0
Towards Watermarking of Open-Source LLMs0
TP-Aware Dequantization0
TQ-DiT: Efficient Time-Aware Quantization for Diffusion Transformers0
Trainable Fixed-Point Quantization for Deep Learning Acceleration on FPGAs0
Train Big, Then Compress: Rethinking Model Size for Efficient Training and Inference of Transformers0
Train Flat, Then Compress: Sharpness-Aware Minimization Learns More Compressible Models0
Training Acceleration of Low-Rank Decomposed Networks using Sequential Freezing and Rank Quantization0
Training and Inference for Integer-Based Semantic Segmentation Network0
Training DNN IoT Applications for Deployment On Analog NVM Crossbars0
Training Integer-Only Deep Recurrent Neural Networks0
Training of mixed-signal optical convolutional neural network with reduced quantization level0
Training Quantized Neural Networks with a Full-precision Auxiliary Module0
Training Quantized Neural Networks to Global Optimality via Semidefinite Programming0
Training with reduced precision of a support vector machine model for text classification0
Transceiver Cooperative Learning-aided Semantic Communications Against Mismatched Background Knowledge Bases0
Transferable Sequential Recommendation via Vector Quantized Meta Learning0
Transfer Hashing with Privileged Information0
Transformations in Learned Image Compression from a Modulation Perspective0
Transform-Based Feature Map Compression for CNN Inference0
Show:102550
← PrevPage 125 of 197Next →

Benchmark Results

#ModelMetricClaimedVerifiedStatus
1FQ-ViT (ViT-L)Top-1 Accuracy (%)85.03Unverified
2FQ-ViT (ViT-B)Top-1 Accuracy (%)83.31Unverified
3FQ-ViT (Swin-B)Top-1 Accuracy (%)82.97Unverified
4FQ-ViT (Swin-S)Top-1 Accuracy (%)82.71Unverified
5FQ-ViT (DeiT-B)Top-1 Accuracy (%)81.2Unverified
6FQ-ViT (Swin-T)Top-1 Accuracy (%)80.51Unverified
7FQ-ViT (DeiT-S)Top-1 Accuracy (%)79.17Unverified
8Xception W8A8Top-1 Accuracy (%)78.97Unverified
9ADLIK-MO-ResNet50-W4A4Top-1 Accuracy (%)77.88Unverified
10ADLIK-MO-ResNet50-W3A4Top-1 Accuracy (%)77.34Unverified
#ModelMetricClaimedVerifiedStatus
13DCNN_VIVA_3MAP160,327.04Unverified
2DTQMAP0.79Unverified
#ModelMetricClaimedVerifiedStatus
1OutEffHop-Bert_basePerplexity6.3Unverified
2OutEffHop-Bert_basePerplexity6.21Unverified
#ModelMetricClaimedVerifiedStatus
1Accuracy98.13Unverified
#ModelMetricClaimedVerifiedStatus
1Accuracy92.92Unverified
#ModelMetricClaimedVerifiedStatus
1SSD ResNet50 V1 FPN 640x640MAP34.3Unverified
#ModelMetricClaimedVerifiedStatus
1TAR @ FAR=1e-495.13Unverified
#ModelMetricClaimedVerifiedStatus
1TAR @ FAR=1e-496.38Unverified
#ModelMetricClaimedVerifiedStatus
13DCNN_VIVA_5All84,809,664Unverified
#ModelMetricClaimedVerifiedStatus
1Accuracy99.8Unverified