SOTAVerified

Quantization

Quantization is a promising technique to reduce the computation cost of neural network training, which can replace high-cost floating-point numbers (e.g., float32) with low-cost fixed-point numbers (e.g., int8/int16).

Source: Adaptive Precision Training: Quantify Back Propagation in Neural Networks with Fixed-point Numbers

Papers

Showing 37263750 of 4925 papers

TitleStatusHype
Algorithm and VLSI Design for 1-bit Data Detection in Massive MIMO-OFDMCode0
Deep data compression for approximate ultrasonic image formation0
Layer-specific Optimization for Mixed Data Flow with Mixed Precision in FPGA Design for CNN-based Object Detectors0
A new heuristic algorithm for fast k-segmentation0
Transform Quantization for CNN (Convolutional Neural Network) Compression0
Object Detection-Based Variable Quantization Processing0
Heatmap Regression via Randomized RoundingCode1
Scaling Up Deep Neural Network Optimization for Edge Inference0
An Integrated Approach to Produce Robust Models with High EfficiencyCode0
Optimal Quantization for Batch Normalization in Neural Network Deployments and Beyond0
An adaptive random experiment design method for engineering experiment0
GAN Slimming: All-in-One GAN Compression by A Unified Optimization FrameworkCode1
Stochastic Markov Gradient Descent and Training Low-Bit Neural Networks0
Convergence of Federated Learning over a Noisy Downlink0
Stochastic Hybrid Combining Design for Quantized Massive MIMO Systems0
Lossy Image Compression with Normalizing Flows0
One Weight Bitwidth to Rule Them All0
Training of mixed-signal optical convolutional neural network with reduced quantization level0
Utilizing Explainable AI for Quantization and Pruning of Deep Neural Networks0
Channel-wise Hessian Aware trace-Weighted Quantization of Neural Networks0
False Detection (Positives and Negatives) in Object Detection0
ECG beats classification via online sparse dictionary and time pyramid matchingCode0
Weight Equalizing Shift Scaler-Coupled Post-training Quantization0
Leveraging Automated Mixed-Low-Precision Quantization for tiny edge microcontrollers0
Compression of Deep Learning Models for Text: A Survey0
Show:102550
← PrevPage 150 of 197Next →

Benchmark Results

#ModelMetricClaimedVerifiedStatus
1FQ-ViT (ViT-L)Top-1 Accuracy (%)85.03Unverified
2FQ-ViT (ViT-B)Top-1 Accuracy (%)83.31Unverified
3FQ-ViT (Swin-B)Top-1 Accuracy (%)82.97Unverified
4FQ-ViT (Swin-S)Top-1 Accuracy (%)82.71Unverified
5FQ-ViT (DeiT-B)Top-1 Accuracy (%)81.2Unverified
6FQ-ViT (Swin-T)Top-1 Accuracy (%)80.51Unverified
7FQ-ViT (DeiT-S)Top-1 Accuracy (%)79.17Unverified
8Xception W8A8Top-1 Accuracy (%)78.97Unverified
9ADLIK-MO-ResNet50-W4A4Top-1 Accuracy (%)77.88Unverified
10ADLIK-MO-ResNet50-W3A4Top-1 Accuracy (%)77.34Unverified
#ModelMetricClaimedVerifiedStatus
13DCNN_VIVA_3MAP160,327.04Unverified
2DTQMAP0.79Unverified
#ModelMetricClaimedVerifiedStatus
1OutEffHop-Bert_basePerplexity6.3Unverified
2OutEffHop-Bert_basePerplexity6.21Unverified
#ModelMetricClaimedVerifiedStatus
1Accuracy98.13Unverified
#ModelMetricClaimedVerifiedStatus
1Accuracy92.92Unverified
#ModelMetricClaimedVerifiedStatus
1SSD ResNet50 V1 FPN 640x640MAP34.3Unverified
#ModelMetricClaimedVerifiedStatus
1TAR @ FAR=1e-495.13Unverified
#ModelMetricClaimedVerifiedStatus
1TAR @ FAR=1e-496.38Unverified
#ModelMetricClaimedVerifiedStatus
13DCNN_VIVA_5All84,809,664Unverified
#ModelMetricClaimedVerifiedStatus
1Accuracy99.8Unverified