SOTAVerified

Quantization

Quantization is a promising technique to reduce the computation cost of neural network training, which can replace high-cost floating-point numbers (e.g., float32) with low-cost fixed-point numbers (e.g., int8/int16).

Source: Adaptive Precision Training: Quantify Back Propagation in Neural Networks with Fixed-point Numbers

Papers

Showing 45764600 of 4925 papers

TitleStatusHype
Efficient Text-driven Motion Generation via Latent Consistency TrainingCode0
Column-wise Quantization of Weights and Partial Sums for Accurate and Efficient Compute-In-Memory AcceleratorsCode0
Rapid-INR: Storage Efficient CPU-free DNN Training Using Implicit Neural RepresentationCode0
Optimal Quantization for Matrix MultiplicationCode0
Learning Category Trees for ID-Based Recommendation: Exploring the Power of Differentiable Vector QuantizationCode0
Learning Frequency-Specific Quantization Scaling in VVC for Standard-Compliant Task-driven Image CodingCode0
Efficient statistical classification of satellite measurementsCode0
Efficient Speech Translation through Model Compression and Knowledge DistillationCode0
Optimization of Armv9 architecture general large language model inference performance based on Llama.cppCode0
Aggregated Learning: A Vector-Quantization Approach to Learning Neural Network ClassifiersCode0
Understanding and Improving Knowledge Distillation for Quantization-Aware Training of Large Transformer EncodersCode0
An efficient and straightforward online quantization method for a data stream through remove-birth updatingCode0
An Edge Computing-Based Solution for Real-Time Leaf Disease Classification using Thermal ImagingCode0
Understanding Cache Boundness of ML Operators on ARM ProcessorsCode0
Towards Quantized Model Parallelism for Graph-Augmented MLPs Based on Gradient-Free ADMM FrameworkCode0
When Quantization Affects Confidence of Large Language Models?Code0
Task Vector Quantization for Memory-Efficient Model MergingCode0
Deep Learning Models in Speech Recognition: Measuring GPU Energy Consumption, Impact of Noise and Model Quantization for Edge DeploymentCode0
Learning Convolutional Transforms for Lossy Point Cloud Geometry CompressionCode0
Learning Compression from Limited Unlabeled DataCode0
Deep Learning-Based Quantization of L-Values for Gray-Coded ModulationCode0
Efficient Quantization-Aware Training on Segment Anything Model in Medical Images and Its DeploymentCode0
Optimizing Deep Neural Networks using Safety-Guided Self CompressionCode0
Efficient Online Inference of Vision Transformers by Training-Free TokenizationCode0
Learning compact binary descriptors with unsupervised deep neural networksCode0
Show:102550
← PrevPage 184 of 197Next →

Benchmark Results

#ModelMetricClaimedVerifiedStatus
1FQ-ViT (ViT-L)Top-1 Accuracy (%)85.03Unverified
2FQ-ViT (ViT-B)Top-1 Accuracy (%)83.31Unverified
3FQ-ViT (Swin-B)Top-1 Accuracy (%)82.97Unverified
4FQ-ViT (Swin-S)Top-1 Accuracy (%)82.71Unverified
5FQ-ViT (DeiT-B)Top-1 Accuracy (%)81.2Unverified
6FQ-ViT (Swin-T)Top-1 Accuracy (%)80.51Unverified
7FQ-ViT (DeiT-S)Top-1 Accuracy (%)79.17Unverified
8Xception W8A8Top-1 Accuracy (%)78.97Unverified
9ADLIK-MO-ResNet50-W4A4Top-1 Accuracy (%)77.88Unverified
10ADLIK-MO-ResNet50-W3A4Top-1 Accuracy (%)77.34Unverified
#ModelMetricClaimedVerifiedStatus
13DCNN_VIVA_3MAP160,327.04Unverified
2DTQMAP0.79Unverified
#ModelMetricClaimedVerifiedStatus
1OutEffHop-Bert_basePerplexity6.3Unverified
2OutEffHop-Bert_basePerplexity6.21Unverified
#ModelMetricClaimedVerifiedStatus
1Accuracy98.13Unverified
#ModelMetricClaimedVerifiedStatus
1Accuracy92.92Unverified
#ModelMetricClaimedVerifiedStatus
1SSD ResNet50 V1 FPN 640x640MAP34.3Unverified
#ModelMetricClaimedVerifiedStatus
1TAR @ FAR=1e-495.13Unverified
#ModelMetricClaimedVerifiedStatus
1TAR @ FAR=1e-496.38Unverified
#ModelMetricClaimedVerifiedStatus
13DCNN_VIVA_5All84,809,664Unverified
#ModelMetricClaimedVerifiedStatus
1Accuracy99.8Unverified