SOTAVerified

Quantization

Quantization is a promising technique to reduce the computation cost of neural network training, which can replace high-cost floating-point numbers (e.g., float32) with low-cost fixed-point numbers (e.g., int8/int16).

Source: Adaptive Precision Training: Quantify Back Propagation in Neural Networks with Fixed-point Numbers

Papers

Showing 34513500 of 4925 papers

TitleStatusHype
A Comprehensive Benchmark for Single Image Compression Artifacts Reduction0
A comprehensive review of Binary Neural Network0
A Comprehensive Study on Quantization Techniques for Large Language Models0
A Comprehensive Survey of Compression Algorithms for Language Models0
A Comprehensive Survey on Model Quantization for Deep Neural Networks in Image Classification0
A Comprehensive Survey on Vector Database: Storage and Retrieval Technique, Challenge0
A Compressed Sensing Approach for Distribution Matching0
A Cost-Efficient FPGA Implementation of Tiny Transformer Model using Neural ODE0
A Counterexample in Cross-Correlation Template Matching0
Acoustic Model Compression with MAP adaptation0
ACQ: Improving Generative Data-free Quantization Via Attention Correction0
ACT360: An Efficient 360-Degree Action Detection and Summarization Framework for Mission-Critical Training and Debriefing0
Action-Quantized Offline Reinforcement Learning for Robotic Skill Learning0
Activation Density based Mixed-Precision Quantization for Energy Efficient Neural Networks0
Activation Functions for Generalized Learning Vector Quantization - A Performance Comparison0
Activation Map-based Vector Quantization for 360-degree Image Semantic Communication0
AdaComp : Adaptive Residual Gradient Compression for Data-Parallel Distributed Training0
Adaptation of MobileNetV2 for Face Detection on Ultra-Low Power Platform0
ADaPTION: Toolbox and Benchmark for Training Convolutional Neural Networks with Reduced Numerical Precision Weights and Activation0
Adaptive Asymmetric Label-guided Hashing for Multimedia Search0
Adaptive Block Floating-Point for Analog Deep Learning Hardware0
Adaptive Compression for Communication-Efficient Distributed Training0
Adaptive Dataset Quantization0
Adaptive Discrete Communication Bottlenecks with Dynamic Vector Quantization0
Adaptive Dithering Using Curved Markov-Gaussian Noise in the Quantized Domain for Mapping SDR to HDR Image0
Adaptive Dither Voting for Robust Spatial Verification0
Adaptive Integrate-and-Fire Time Encoding Machine with Quantization0
Adaptive Joint Optimization for 3D Reconstruction with Differentiable Rendering0
Adaptive Low-Precision Training for Embeddings in Click-Through Rate Prediction0
Adaptive Periodic Averaging: A Practical Approach to Reducing Communication in Distributed Learning0
Adaptive Precision Training: Quantify Back Propagation in Neural Networks with Fixed-point Numbers0
Adaptive Proximal Gradient Methods for Structured Neural Networks0
Adaptive Quantization for Deep Neural Network0
Adaptive Quantization for Key Generation in Low-Power Wide-Area Networks0
Adaptive Quantization of Model Updates for Communication-Efficient Federated Learning0
Adaptive Quantization of Neural Networks0
Adaptive Quantization Resolution and Power Control for Federated Learning over Cell-free Networks0
Adaptive quantization with mixed-precision based on low-cost proxy0
Adaptive Resolution Inference (ARI): Energy-Efficient Machine Learning for Internet of Things0
Adaptive Resource Allocation for Semantic Communication Networks0
Adaptive Sample-space & Adaptive Probability coding: a neural-network based approach for compression0
Adaptive Training of Random Mapping for Data Quantization0
Adaptive Transmission for Distributed Detection in Energy Harvesting Wireless Sensor Networks0
Adaptive Wireless Image Semantic Transmission and Over-The-Air Testing0
Adaptive Wireless Image Semantic Transmission: Design, Simulation, and Prototype Validation0
AdaptivFloat: A Floating-point based Data Type for Resilient Deep Learning Inference0
AdaQAT: Adaptive Bit-Width Quantization-Aware Training0
A Data and Compute Efficient Design for Limited-Resources Deep Learning0
AdderNet and its Minimalist Hardware Design for Energy-Efficient Artificial Intelligence0
Additive Quantization for Extreme Vector Compression0
Show:102550
← PrevPage 70 of 99Next →

Benchmark Results

#ModelMetricClaimedVerifiedStatus
1FQ-ViT (ViT-L)Top-1 Accuracy (%)85.03Unverified
2FQ-ViT (ViT-B)Top-1 Accuracy (%)83.31Unverified
3FQ-ViT (Swin-B)Top-1 Accuracy (%)82.97Unverified
4FQ-ViT (Swin-S)Top-1 Accuracy (%)82.71Unverified
5FQ-ViT (DeiT-B)Top-1 Accuracy (%)81.2Unverified
6FQ-ViT (Swin-T)Top-1 Accuracy (%)80.51Unverified
7FQ-ViT (DeiT-S)Top-1 Accuracy (%)79.17Unverified
8Xception W8A8Top-1 Accuracy (%)78.97Unverified
9ADLIK-MO-ResNet50-W4A4Top-1 Accuracy (%)77.88Unverified
10ADLIK-MO-ResNet50-W3A4Top-1 Accuracy (%)77.34Unverified
#ModelMetricClaimedVerifiedStatus
13DCNN_VIVA_3MAP160,327.04Unverified
2DTQMAP0.79Unverified
#ModelMetricClaimedVerifiedStatus
1OutEffHop-Bert_basePerplexity6.3Unverified
2OutEffHop-Bert_basePerplexity6.21Unverified
#ModelMetricClaimedVerifiedStatus
1Accuracy98.13Unverified
#ModelMetricClaimedVerifiedStatus
1Accuracy92.92Unverified
#ModelMetricClaimedVerifiedStatus
1SSD ResNet50 V1 FPN 640x640MAP34.3Unverified
#ModelMetricClaimedVerifiedStatus
1TAR @ FAR=1e-495.13Unverified
#ModelMetricClaimedVerifiedStatus
1TAR @ FAR=1e-496.38Unverified
#ModelMetricClaimedVerifiedStatus
13DCNN_VIVA_5All84,809,664Unverified
#ModelMetricClaimedVerifiedStatus
1Accuracy99.8Unverified