SOTAVerified

Quantization

Quantization is a promising technique to reduce the computation cost of neural network training, which can replace high-cost floating-point numbers (e.g., float32) with low-cost fixed-point numbers (e.g., int8/int16).

Source: Adaptive Precision Training: Quantify Back Propagation in Neural Networks with Fixed-point Numbers

Papers

Showing 21012150 of 4925 papers

TitleStatusHype
FlashEval: Towards Fast and Accurate Evaluation of Text-to-image Diffusion Generative ModelsCode0
SignSGD with Federated VotingCode0
Work in Progress: Linear Transformers for TinyML0
On the Impact of Black-box Deployment Strategies for Edge AI on Latency and Model PerformanceCode0
Neural Image Compression with Quantization Rectifier0
GLAD: Improving Latent Graph Generative Modeling with Simple QuantizationCode0
Infrastructure-Assisted Collaborative Perception in Automated Valet Parking: A Safety Perspective0
Magic for the Age of Quantized DNNs0
Super-High-Fidelity Image Compression via Hierarchical-ROI and Adaptive Quantization0
NoisyDECOLLE: Robust Local Learning for SNNs on Neuromorphic HardwareCode0
Provable Privacy with Non-Private Pre-Processing0
Floating-Point Quantization Analysis of Multi-Layer Perceptron Artificial Neural NetworksCode0
HyperVQ: MLR-based Vector Quantization in Hyperbolic Space0
Spatio-Temporal Fluid Dynamics Modeling via Physical-Awareness and Parameter Diffusion Guidance0
Decoding Compressed Trust: Scrutinizing the Trustworthiness of Efficient LLMs Under Compression0
Hierarchical Frequency-based Upsampling and Refining for Compressed Video Quality Enhancement0
Quantization Avoids Saddle Points in Distributed Optimization0
Quantization Effects on Neural Networks Perception: How would quantization change the perceptual field of vision models?Code0
CRB Analysis for Mixed-ADC Based DOA Estimation0
Generalized Relevance Learning Grassmann QuantizationCode0
Adversarial Fine-tuning of Compressed Neural Networks for Joint Improvement of Robustness and EfficiencyCode0
FedComLoc: Communication-Efficient Distributed Training of Sparse and Quantized Models0
UniCode: Learning a Unified Codebook for Multimodal Large Language Models0
BRIEDGE: EEG-Adaptive Edge AI for Multi-Brain to Multi-Robot Interaction0
Strategizing against Q-learners: A Control-theoretical Approach0
Collaborative Automotive Radar Sensing via Mixed-Precision Distributed Array Completion0
Vector Quantization for Deep-Learning-Based CSI Feedback in Massive MIMO Systems0
Approaching Rate-Distortion Limits in Neural Compression with Lattice Transform Coding0
FlowVQTalker: High-Quality Emotional Talking Face Generation through Normalizing Flow and Quantization0
What Makes Quantization for Large Language Models Hard? An Empirical Study from the Lens of Perturbation0
QuantTune: Optimizing Model Quantization with Adaptive Outlier-Driven Fine Tuning0
Enhancing Multimodal Unified Representations for Cross Modal Generalization0
Micro-Fracture Detection in Photovoltaic Cells with Hardware-Constrained Devices and Computer Vision0
The Impact of Quantization on the Robustness of Transformer-based Text Classifiers0
Algorithm-Hardware Co-Design of Distribution-Aware Logarithmic-Posit Encodings for Efficient DNN InferenceCode0
LoCoDL: Communication-Efficient Distributed Learning with Local Training and Compression0
On-demand Quantization for Green Federated Generative Diffusion in Mobile Edge Networks0
Adaptive Integrate-and-Fire Time Encoding Machine with Quantization0
EasyQuant: An Efficient Data-free Quantization Algorithm for LLMs0
Design of Stochastic Quantizers for Privacy Preservation0
VQSynery: Robust Drug Synergy Prediction With Vector Quantization Mechanism0
Deep-Learned Compression for Radio-Frequency Signal Classification0
FlowPrecision: Advancing FPGA-Based Real-Time Fluid Flow Estimation with Linear Quantization0
Neural Network Assisted Lifting Steps For Improved Fully Scalable Lossy Image Compression in JPEG 2000Code0
Towards efficient deep autoencoders for multivariate time series anomaly detection0
Better Schedules for Low Precision Training of Deep Neural Networks0
A Hierarchical Federated Learning Approach for the Internet of Things0
On the Compressibility of Quantized Large Language Models0
Extracting Usable Predictions from Quantized Networks through Uncertainty Quantification for OOD DetectionCode0
BasedAI: A decentralized P2P network for Zero Knowledge Large Language Models (ZK-LLMs)0
Show:102550
← PrevPage 43 of 99Next →

Benchmark Results

#ModelMetricClaimedVerifiedStatus
1FQ-ViT (ViT-L)Top-1 Accuracy (%)85.03Unverified
2FQ-ViT (ViT-B)Top-1 Accuracy (%)83.31Unverified
3FQ-ViT (Swin-B)Top-1 Accuracy (%)82.97Unverified
4FQ-ViT (Swin-S)Top-1 Accuracy (%)82.71Unverified
5FQ-ViT (DeiT-B)Top-1 Accuracy (%)81.2Unverified
6FQ-ViT (Swin-T)Top-1 Accuracy (%)80.51Unverified
7FQ-ViT (DeiT-S)Top-1 Accuracy (%)79.17Unverified
8Xception W8A8Top-1 Accuracy (%)78.97Unverified
9ADLIK-MO-ResNet50-W4A4Top-1 Accuracy (%)77.88Unverified
10ADLIK-MO-ResNet50-W3A4Top-1 Accuracy (%)77.34Unverified
#ModelMetricClaimedVerifiedStatus
13DCNN_VIVA_3MAP160,327.04Unverified
2DTQMAP0.79Unverified
#ModelMetricClaimedVerifiedStatus
1OutEffHop-Bert_basePerplexity6.3Unverified
2OutEffHop-Bert_basePerplexity6.21Unverified
#ModelMetricClaimedVerifiedStatus
1Accuracy98.13Unverified
#ModelMetricClaimedVerifiedStatus
1Accuracy92.92Unverified
#ModelMetricClaimedVerifiedStatus
1SSD ResNet50 V1 FPN 640x640MAP34.3Unverified
#ModelMetricClaimedVerifiedStatus
1TAR @ FAR=1e-495.13Unverified
#ModelMetricClaimedVerifiedStatus
1TAR @ FAR=1e-496.38Unverified
#ModelMetricClaimedVerifiedStatus
13DCNN_VIVA_5All84,809,664Unverified
#ModelMetricClaimedVerifiedStatus
1Accuracy99.8Unverified