SOTAVerified

Quantization

Quantization is a promising technique to reduce the computation cost of neural network training, which can replace high-cost floating-point numbers (e.g., float32) with low-cost fixed-point numbers (e.g., int8/int16).

Source: Adaptive Precision Training: Quantify Back Propagation in Neural Networks with Fixed-point Numbers

Papers

Showing 351400 of 4925 papers

TitleStatusHype
Robust Machine Unlearning for Quantized Neural Networks via Adaptive Gradient Reweighting with Similar Labels0
Quantization-Free Autoregressive Action TransformerCode0
MAG: Multi-Modal Aligned Autoregressive Co-Speech Gesture Generation without Vector Quantization0
CompMarkGS: Robust Watermarking for Compressed 3D Gaussian Splatting0
ClusComp: A Simple Paradigm for Model Compression and Efficient Finetuning0
ML-SpecQD: Multi-Level Speculative Decoding with Quantized Drafts0
Versatile Physics-based Character Control with Hybrid Latent Representation0
ACT360: An Efficient 360-Degree Action Detection and Summarization Framework for Mission-Critical Training and Debriefing0
Stabilizing Quantization-Aware Training by Implicit-Regularization on Hessian Matrix0
Understanding Flatness in Generative Models: Its Role and Benefits0
Pathology Image Compression with Pre-trained Autoencoders0
OuroMamba: A Data-Free Quantization Framework for Vision Mamba Models0
Automated Tomato Maturity Estimation Using an Optimized Residual Model with Pruning and Quantization Techniques0
Dual Codebook VQ: Enhanced Image Reconstruction with Reduced Codebook Size0
Global synchronization of multi-agent systems with nonlinear interactions0
Quantization for OpenAI's Whisper Models: A Comparative AnalysisCode0
ViM-VQ: Efficient Post-Training Vector Quantization for Visual Mamba0
Sometimes Painful but Certainly Promising: Feasibility and Trade-offs of Language Model Inference at the Edge0
Quantitative Analysis of Deeply Quantized Tiny Neural Networks Robust to Adversarial Attacks0
Accurate INT8 Training Through Dynamic Block-Level Fallback0
PCGS: Progressive Compression of 3D Gaussian SplattingCode1
Quantization Design for Deep Learning-Based CSI Feedback0
PRISM: Privacy-Preserving Improved Stochastic Masking for Federated Generative ModelsCode0
VocalEyes: Enhancing Environmental Perception for the Visually Impaired through Vision-Language Models and Distance-Aware Object Detection0
Non-vacuous Generalization Bounds for Deep Neural Networks without any modification to the trained models0
Post-Training Quantization for Diffusion Transformer via Hierarchical Timestep Grouping0
Breaking the Limits of Quantization-Aware Defenses: QADT-R for Robustness Against Patch-Based Adversarial Attacks in QNNs0
Task Vector Quantization for Memory-Efficient Model MergingCode0
Synchronized Video-to-Audio Generation via Mel Quantization-Continuum Decomposition0
Lightweight Multimodal Artificial Intelligence Framework for Maritime Multi-Scene Recognition0
QuantCache: Adaptive Importance-Guided Quantization with Hierarchical Latent and Layer Caching for Video GenerationCode1
TR-DQ: Time-Rotation Diffusion Quantization0
PathVQ: Reforming Computational Pathology Foundation Model for Whole Slide Image Analysis via Vector Quantization0
Towards Superior Quantization Accuracy: A Layer-sensitive Approach0
SAQ-SAM: Semantically-Aligned Quantization for Segment Anything Model0
Does Acceleration Cause Hidden Instability in Vision Language Models? Uncovering Instance-Level Divergence Through a Large-Scale Empirical Study0
Seeing Delta Parameters as JPEG Images: Data-Free Delta Compression with Discrete Cosine Transform0
D2GV: Deformable 2D Gaussian Splatting for Video Representation in 400FPSCode2
QArtSR: Quantization via Reverse-Module and Timestep-Retraining in One-Step Diffusion based Image Super-ResolutionCode1
CASP: Compression of Large Multimodal Models Based on Attention SparsityCode0
Frequency Autoregressive Image Generation with Continuous Tokens0
Discrete Contrastive Learning for Diffusion Policies in Autonomous Driving0
VQEL: Enabling Self-Developed Symbolic Language in Agents through Vector Quantization in Emergent Language Games0
LVLM-Compress-Bench: Benchmarking the Broader Impact of Large Vision-Language Model CompressionCode0
End-to-End Human Pose Reconstruction from Wearable Sensors for 6G Extended Reality SystemsCode0
Universality of Layer-Level Entropy-Weighted Quantization Beyond Model Architecture and Size0
Lightweight Embedded FPGA Deployment of Learned Image Compression with Knowledge Distillation and Hybrid Quantization0
On the Relation Between Speech Quality and Quantized Latent Representations of Neural Codecs0
English K_Quantization of LLMs Does Not Disproportionately Diminish Multilingual Performance0
Fast Jet Tagging with MLP-Mixers on FPGAs0
Show:102550
← PrevPage 8 of 99Next →

Benchmark Results

#ModelMetricClaimedVerifiedStatus
1FQ-ViT (ViT-L)Top-1 Accuracy (%)85.03Unverified
2FQ-ViT (ViT-B)Top-1 Accuracy (%)83.31Unverified
3FQ-ViT (Swin-B)Top-1 Accuracy (%)82.97Unverified
4FQ-ViT (Swin-S)Top-1 Accuracy (%)82.71Unverified
5FQ-ViT (DeiT-B)Top-1 Accuracy (%)81.2Unverified
6FQ-ViT (Swin-T)Top-1 Accuracy (%)80.51Unverified
7FQ-ViT (DeiT-S)Top-1 Accuracy (%)79.17Unverified
8Xception W8A8Top-1 Accuracy (%)78.97Unverified
9ADLIK-MO-ResNet50-W4A4Top-1 Accuracy (%)77.88Unverified
10ADLIK-MO-ResNet50-W3A4Top-1 Accuracy (%)77.34Unverified
#ModelMetricClaimedVerifiedStatus
13DCNN_VIVA_3MAP160,327.04Unverified
2DTQMAP0.79Unverified
#ModelMetricClaimedVerifiedStatus
1OutEffHop-Bert_basePerplexity6.3Unverified
2OutEffHop-Bert_basePerplexity6.21Unverified
#ModelMetricClaimedVerifiedStatus
1Accuracy98.13Unverified
#ModelMetricClaimedVerifiedStatus
1Accuracy92.92Unverified
#ModelMetricClaimedVerifiedStatus
1SSD ResNet50 V1 FPN 640x640MAP34.3Unverified
#ModelMetricClaimedVerifiedStatus
1TAR @ FAR=1e-495.13Unverified
#ModelMetricClaimedVerifiedStatus
1TAR @ FAR=1e-496.38Unverified
#ModelMetricClaimedVerifiedStatus
13DCNN_VIVA_5All84,809,664Unverified
#ModelMetricClaimedVerifiedStatus
1Accuracy99.8Unverified