SOTAVerified

Quantization

Quantization is a promising technique to reduce the computation cost of neural network training, which can replace high-cost floating-point numbers (e.g., float32) with low-cost fixed-point numbers (e.g., int8/int16).

Source: Adaptive Precision Training: Quantify Back Propagation in Neural Networks with Fixed-point Numbers

Papers

Showing 45514600 of 4925 papers

TitleStatusHype
Deep Metric Learning to RankCode0
Learning to Seek: Autonomous Source Seeking with Deep Reinforcement Learning Onboard a Nano Drone MicrocontrollerCode0
Layer-Wise Quantization: A Pragmatic and Effective Method for Quantizing LLMs Beyond Integer Bit-LevelsCode0
AdaBin: Improving Binary Neural Networks with Adaptive Binary SetsCode0
Empirical Evaluation of Deep Learning Model Compression Techniques on the WaveNet VocoderCode0
PLUM: Improving Inference Efficiency By Leveraging Repetition-Sparsity Trade-OffCode0
SignSGD with Federated Defense: Harnessing Adversarial Attacks through Gradient Sign DecodingCode0
Open-source FPGA-ML codesign for the MLPerf Tiny BenchmarkCode0
Learning Space Partitions for Nearest Neighbor SearchCode0
Random and Adversarial Bit Error Robustness: Energy-Efficient and Secure DNN AcceleratorsCode0
Operations Guided Neural Networks for High Fidelity Data-To-Text GenerationCode0
SignSGD with Federated VotingCode0
Random Entity Quantization for Parameter-Efficient Compositional Knowledge Graph RepresentationCode0
EmbBERT-Q: Breaking Memory Barriers in Embedded NLPCode0
Randomized Quantization is All You Need for Differential Privacy in Federated LearningCode0
Automatic Neural Network Compression by Sparsity-Quantization Joint Learning: A Constrained Optimization-based ApproachCode0
Elastic Product Quantization for Time SeriesCode0
Learning Semantic Textual Similarity via Topic-informed Discrete Latent VariablesCode0
Optimal Clipping and Magnitude-aware Differentiation for Improved Quantization-aware TrainingCode0
Task-Based Graph Signal CompressionCode0
Unconditional Image-Text Pair Generation with Multimodal Cross QuantizerCode0
Deep Log-Likelihood Ratio QuantizationCode0
Learning Physical-Layer Communication with Quantized FeedbackCode0
Deep Learning with Low Precision by Half-wave Gaussian QuantizationCode0
Efficiera Residual Networks: Hardware-Friendly Fully Binary Weight with 2-bit Activation Model Achieves Practical ImageNet AccuracyCode0
Efficient Text-driven Motion Generation via Latent Consistency TrainingCode0
Column-wise Quantization of Weights and Partial Sums for Accurate and Efficient Compute-In-Memory AcceleratorsCode0
Rapid-INR: Storage Efficient CPU-free DNN Training Using Implicit Neural RepresentationCode0
Optimal Quantization for Matrix MultiplicationCode0
Learning Category Trees for ID-Based Recommendation: Exploring the Power of Differentiable Vector QuantizationCode0
Learning Frequency-Specific Quantization Scaling in VVC for Standard-Compliant Task-driven Image CodingCode0
Efficient statistical classification of satellite measurementsCode0
Efficient Speech Translation through Model Compression and Knowledge DistillationCode0
Optimization of Armv9 architecture general large language model inference performance based on Llama.cppCode0
Aggregated Learning: A Vector-Quantization Approach to Learning Neural Network ClassifiersCode0
Understanding and Improving Knowledge Distillation for Quantization-Aware Training of Large Transformer EncodersCode0
An efficient and straightforward online quantization method for a data stream through remove-birth updatingCode0
An Edge Computing-Based Solution for Real-Time Leaf Disease Classification using Thermal ImagingCode0
Understanding Cache Boundness of ML Operators on ARM ProcessorsCode0
Towards Quantized Model Parallelism for Graph-Augmented MLPs Based on Gradient-Free ADMM FrameworkCode0
When Quantization Affects Confidence of Large Language Models?Code0
Task Vector Quantization for Memory-Efficient Model MergingCode0
Deep Learning Models in Speech Recognition: Measuring GPU Energy Consumption, Impact of Noise and Model Quantization for Edge DeploymentCode0
Learning Convolutional Transforms for Lossy Point Cloud Geometry CompressionCode0
Learning Compression from Limited Unlabeled DataCode0
Deep Learning-Based Quantization of L-Values for Gray-Coded ModulationCode0
Efficient Quantization-Aware Training on Segment Anything Model in Medical Images and Its DeploymentCode0
Optimizing Deep Neural Networks using Safety-Guided Self CompressionCode0
Efficient Online Inference of Vision Transformers by Training-Free TokenizationCode0
Learning compact binary descriptors with unsupervised deep neural networksCode0
Show:102550
← PrevPage 92 of 99Next →

Benchmark Results

#ModelMetricClaimedVerifiedStatus
1FQ-ViT (ViT-L)Top-1 Accuracy (%)85.03Unverified
2FQ-ViT (ViT-B)Top-1 Accuracy (%)83.31Unverified
3FQ-ViT (Swin-B)Top-1 Accuracy (%)82.97Unverified
4FQ-ViT (Swin-S)Top-1 Accuracy (%)82.71Unverified
5FQ-ViT (DeiT-B)Top-1 Accuracy (%)81.2Unverified
6FQ-ViT (Swin-T)Top-1 Accuracy (%)80.51Unverified
7FQ-ViT (DeiT-S)Top-1 Accuracy (%)79.17Unverified
8Xception W8A8Top-1 Accuracy (%)78.97Unverified
9ADLIK-MO-ResNet50-W4A4Top-1 Accuracy (%)77.88Unverified
10ADLIK-MO-ResNet50-W3A4Top-1 Accuracy (%)77.34Unverified
#ModelMetricClaimedVerifiedStatus
13DCNN_VIVA_3MAP160,327.04Unverified
2DTQMAP0.79Unverified
#ModelMetricClaimedVerifiedStatus
1OutEffHop-Bert_basePerplexity6.3Unverified
2OutEffHop-Bert_basePerplexity6.21Unverified
#ModelMetricClaimedVerifiedStatus
1Accuracy98.13Unverified
#ModelMetricClaimedVerifiedStatus
1Accuracy92.92Unverified
#ModelMetricClaimedVerifiedStatus
1SSD ResNet50 V1 FPN 640x640MAP34.3Unverified
#ModelMetricClaimedVerifiedStatus
1TAR @ FAR=1e-495.13Unverified
#ModelMetricClaimedVerifiedStatus
1TAR @ FAR=1e-496.38Unverified
#ModelMetricClaimedVerifiedStatus
13DCNN_VIVA_5All84,809,664Unverified
#ModelMetricClaimedVerifiedStatus
1Accuracy99.8Unverified