SOTAVerified

Quantization

Quantization is a promising technique to reduce the computation cost of neural network training, which can replace high-cost floating-point numbers (e.g., float32) with low-cost fixed-point numbers (e.g., int8/int16).

Source: Adaptive Precision Training: Quantify Back Propagation in Neural Networks with Fixed-point Numbers

Papers

Showing 24012425 of 4925 papers

TitleStatusHype
Efficient and Effective Methods for Mixed Precision Neural Network Quantization for Faster, Energy-efficient Inference0
Parsimonious System Identification from Fragmented Quantized Measurements0
State Machine-based Waveforms for Channels With 1-Bit Quantization and Oversampling With Time-Instance Zero-Crossing Modulation0
BAFFLE: A Baseline of Backpropagation-Free Federated LearningCode1
Understanding INT4 Quantization for Transformer Models: Latency Speedup, Composability, and Failure Cases0
BOMP-NAS: Bayesian Optimization Mixed Precision NAS0
Quantized Deep Path-following Control on a Microcontroller0
Hardware Implementation of Task-based Quantization in Multi-user Signal Recovery0
PowerQuant: Automorphism Search for Non-Uniform Quantization0
Victoria Amazonica Optimization (VAO): An Algorithm Inspired by the Giant Water Lily PlantCode0
RawHash: Enabling Fast and Accurate Real-Time Analysis of Raw Nanopore Signals for Large GenomesCode1
Accelerating and Compressing Deep Neural Networks for Massive MIMO CSI FeedbackCode0
Optimized learned entropy coding parameters for practical neural-based image and video compression0
Robust Symbol Level Precoding for Overlay Cognitive Radio Networks0
Learning Frequency-Specific Quantization Scaling in VVC for Standard-Compliant Task-driven Image CodingCode0
HCE: Improving Performance and Efficiency with Heterogeneously Compressed Neural Network Ensemble0
ACQ: Improving Generative Data-free Quantization Via Attention Correction0
Development, Optimization, and Deployment of Thermal Forward Vision Systems for Advance Vehicular Applications on Edge DevicesCode0
Deep Conditional Measure Quantization0
Masked Vector Quantization0
RedBit: An End-to-End Flexible Framework for Evaluating the Accuracy of Quantized CNNsCode0
Semantic and Effective Communication for Remote Control Tasks with Dynamic Feature Compression0
Exploring Automatic Gym Workouts Recognition Locally On Wearable Resource-Constrained Devices0
Security-Aware Approximate Spiking Neural Networks0
UnifySpeech: A Unified Framework for Zero-shot Text-to-Speech and Voice Conversion0
Show:102550
← PrevPage 97 of 197Next →

Benchmark Results

#ModelMetricClaimedVerifiedStatus
1FQ-ViT (ViT-L)Top-1 Accuracy (%)85.03Unverified
2FQ-ViT (ViT-B)Top-1 Accuracy (%)83.31Unverified
3FQ-ViT (Swin-B)Top-1 Accuracy (%)82.97Unverified
4FQ-ViT (Swin-S)Top-1 Accuracy (%)82.71Unverified
5FQ-ViT (DeiT-B)Top-1 Accuracy (%)81.2Unverified
6FQ-ViT (Swin-T)Top-1 Accuracy (%)80.51Unverified
7FQ-ViT (DeiT-S)Top-1 Accuracy (%)79.17Unverified
8Xception W8A8Top-1 Accuracy (%)78.97Unverified
9ADLIK-MO-ResNet50-W4A4Top-1 Accuracy (%)77.88Unverified
10ADLIK-MO-ResNet50-W3A4Top-1 Accuracy (%)77.34Unverified
#ModelMetricClaimedVerifiedStatus
13DCNN_VIVA_3MAP160,327.04Unverified
2DTQMAP0.79Unverified
#ModelMetricClaimedVerifiedStatus
1OutEffHop-Bert_basePerplexity6.3Unverified
2OutEffHop-Bert_basePerplexity6.21Unverified
#ModelMetricClaimedVerifiedStatus
1Accuracy98.13Unverified
#ModelMetricClaimedVerifiedStatus
1Accuracy92.92Unverified
#ModelMetricClaimedVerifiedStatus
1SSD ResNet50 V1 FPN 640x640MAP34.3Unverified
#ModelMetricClaimedVerifiedStatus
1TAR @ FAR=1e-495.13Unverified
#ModelMetricClaimedVerifiedStatus
1TAR @ FAR=1e-496.38Unverified
#ModelMetricClaimedVerifiedStatus
13DCNN_VIVA_5All84,809,664Unverified
#ModelMetricClaimedVerifiedStatus
1Accuracy99.8Unverified