SOTAVerified

Quantization

Quantization is a promising technique to reduce the computation cost of neural network training, which can replace high-cost floating-point numbers (e.g., float32) with low-cost fixed-point numbers (e.g., int8/int16).

Source: Adaptive Precision Training: Quantify Back Propagation in Neural Networks with Fixed-point Numbers

Papers

Showing 17011750 of 4925 papers

TitleStatusHype
Thinking in Granularity: Dynamic Quantization for Image Super-Resolution by Intriguing Multi-Granularity CluesCode0
CorBin-FL: A Differentially Private Federated Learning Mechanism using Common Randomness0
Reduced bit median quantization: A middle process for Efficient Image Compression0
PTQ4ADM: Post-Training Quantization for Efficient Text Conditional Audio Diffusion Models0
TalkMosaic: Interactive PhotoMosaic with Multi-modal LLM Q&A Interactions0
Impact of ML Optimization Tactics on Greener Pre-Trained ML Models0
NDVQ: Robust Neural Audio Codec with Normal Distribution-Based Vector Quantization0
Scaling FP8 training to trillion-token LLMs0
Art and Science of Quantizing Large-Scale Models: A Comprehensive Overview0
Pareto Data Framework: Steps Towards Resource-Efficient Decision Making Using Minimum Viable Data (MVD)0
Low Frame-rate Speech Codec: a Codec Designed for Fast High-quality Speech LLM Training and Inference0
Exploring the Trade-Offs: Quantization Methods, Task Difficulty, and Model Size in Large Language Models From Edge to GiantCode0
LASERS: LAtent Space Encoding for Representations with Sparsity for Generative Modeling0
Forearm Ultrasound based Gesture Recognition on Edge0
Improving Statistical Significance in Human Evaluation of Automatic Metrics via Soft Pairwise Accuracy0
Language Models and Retrieval Augmented Generation for Automated Structured Data Extraction from Diagnostic Reports0
MesonGS: Post-training Compression of 3D Gaussians via Efficient Attribute Transformation0
Robust Training of Neural Networks at Arbitrary Precision and Sparsity0
Privacy-Preserving SAM Quantization for Efficient Edge Intelligence in Healthcare0
Investigating Disentanglement in a Phoneme-level Speech Codec for Prosody Modeling0
Efficient and Reliable Vector Similarity Search Using Asymmetric Encoding with NAND-Flash for Many-Class Few-Shot Learning0
Dequantization of a signal from two parallel quantized observations0
Adaptive Error-Bounded Hierarchical Matrices for Efficient Neural Network Compression0
Distributed Convolutional Neural Network Training on Mobile and Edge Clusters0
NVRC: Neural Video Representation Compression0
STORE: Streamlining Semantic Tokenization and Generative Recommendation with A Single LLM0
AgileIR: Memory-Efficient Group Shifted Windows Attention for Agile Image Restoration0
Rate-Constrained Quantization for Communication-Efficient Federated Learning0
Estimating the Completeness of Discrete Speech Units0
ECG Biometric Authentication Using Self-Supervised Learning for IoT Edge Sensors0
TriplePlay: Enhancing Federated Learning with CLIP for Non-IID Data and Resource Efficiency0
Distributed Optimization with Finite Bit Adaptive Quantization for Efficient Communication and Precision Enhancement0
SGC-VQGAN: Towards Complex Scene Representation via Semantic Guided Clustering Codebook0
Blind-Adaptive Quantizers0
OPAL: Outlier-Preserved Microscaling Quantization Accelerator for Generative Large Language Models0
Investigating Privacy Bias in Training Data of Language Models0
WaterMAS: Sharpness-Aware Maximization for Neural Network Watermarking0
LAST: Language Model Aware Speech Tokenization0
Recursive Quantization for L_2 Stabilization of a Finite Capacity Stochastic Control Loop with Intermittent State Observations0
CoAst: Validation-Free Contribution Assessment for Federated Learning based on Cross-Round Valuation0
Gaussian Rate-Distortion-Perception Coding and Entropy-Constrained Scalar Quantization0
Task-Oriented Communication for Graph Data: A Graph Information Bottleneck Approach0
Learning Task-Based Trainable Neuromorphic ADCs via Power-Aware Distillation0
Sorbet: A Neuromorphic Hardware-Compatible Transformer-Based Spiking Language Model0
Foundations of Large Language Model Compression -- Part 1: Weight QuantizationCode0
Optimization and Deployment of Deep Neural Networks for PPG-based Blood Pressure Estimation Targeting Low-power Wearables0
Robust Clustering on High-Dimensional Data with Stochastic QuantizationCode0
Compressing VAE-Based Out-of-Distribution Detectors for Embedded Deployment0
One-Index Vector Quantization Based Adversarial Attack on Image Classification0
Edge AI: Evaluation of Model Compression Techniques for Convolutional Neural Networks0
Show:102550
← PrevPage 35 of 99Next →

Benchmark Results

#ModelMetricClaimedVerifiedStatus
1FQ-ViT (ViT-L)Top-1 Accuracy (%)85.03Unverified
2FQ-ViT (ViT-B)Top-1 Accuracy (%)83.31Unverified
3FQ-ViT (Swin-B)Top-1 Accuracy (%)82.97Unverified
4FQ-ViT (Swin-S)Top-1 Accuracy (%)82.71Unverified
5FQ-ViT (DeiT-B)Top-1 Accuracy (%)81.2Unverified
6FQ-ViT (Swin-T)Top-1 Accuracy (%)80.51Unverified
7FQ-ViT (DeiT-S)Top-1 Accuracy (%)79.17Unverified
8Xception W8A8Top-1 Accuracy (%)78.97Unverified
9ADLIK-MO-ResNet50-W4A4Top-1 Accuracy (%)77.88Unverified
10ADLIK-MO-ResNet50-W3A4Top-1 Accuracy (%)77.34Unverified
#ModelMetricClaimedVerifiedStatus
13DCNN_VIVA_3MAP160,327.04Unverified
2DTQMAP0.79Unverified
#ModelMetricClaimedVerifiedStatus
1OutEffHop-Bert_basePerplexity6.3Unverified
2OutEffHop-Bert_basePerplexity6.21Unverified
#ModelMetricClaimedVerifiedStatus
1Accuracy98.13Unverified
#ModelMetricClaimedVerifiedStatus
1Accuracy92.92Unverified
#ModelMetricClaimedVerifiedStatus
1SSD ResNet50 V1 FPN 640x640MAP34.3Unverified
#ModelMetricClaimedVerifiedStatus
1TAR @ FAR=1e-495.13Unverified
#ModelMetricClaimedVerifiedStatus
1TAR @ FAR=1e-496.38Unverified
#ModelMetricClaimedVerifiedStatus
13DCNN_VIVA_5All84,809,664Unverified
#ModelMetricClaimedVerifiedStatus
1Accuracy99.8Unverified