SOTAVerified

Quantization

Quantization is a promising technique to reduce the computation cost of neural network training, which can replace high-cost floating-point numbers (e.g., float32) with low-cost fixed-point numbers (e.g., int8/int16).

Source: Adaptive Precision Training: Quantify Back Propagation in Neural Networks with Fixed-point Numbers

Papers

Showing 38513900 of 4925 papers

TitleStatusHype
Knowledge Distillation-aided End-to-End Learning for Linear Precoding in Multiuser MIMO Downlink Systems with Finite-Rate Feedback0
Distance-Weighted Graph Neural Networks on FPGAs for Real-Time Particle Reconstruction in High Energy Physics0
In-Depth DCT Coefficient Distribution Analysis for First Quantization Estimation0
Subjective Quality Database and Objective Study of Compressed Point Clouds With 6DoF Head-Mounted Display0
Fast top-K Cosine Similarity Search through XOR-Friendly Binary Quantization on GPUs0
TREND: Transferability based Robust ENsemble DesignCode0
A Novel Physics-based Channel Model for Reconfigurable Intelligent Surface-assisted Multi-user Communication Systems0
Cluster-Based Cooperative Digital Over-the-Air Aggregation for Wireless Federated Edge Learning0
The Rate-Distortion-Accuracy Tradeoff: JPEG Case Study0
Super-relaxation of space-time-quantized ensemble of energy loads to curtail their synchronization after demand response perturbation0
Uplink Achievable Rate of Intelligent Reflecting Surface-Aided Millimeter-Wave Communications with Low-Resolution ADC and Phase Noise0
Deep Multi-modality Soft-decoding of Very Low Bit-rate Face Videos0
High-quality Single-model Deep Video Compression with Frame-Conv3D and Multi-frame Differential Modulation0
FTL: A universal framework for training low-bit DNNs via Feature Transfer0
ProxyBNN: Learning Binarized Neural Networks via Proxy Matrices0
Task-Aware Quantization Network for JPEG Image Compression0
Binarized Neural Network for Single Image Super Resolution0
Automatic Gain Control Design for Dynamic Visible Light Communication Systems0
Communication-Efficient Federated Learning via Optimal Client Sampling0
Flexible framework for audio reconstructionCode0
On the Impact of Lossy Image and Video Compression on the Performance of Deep Convolutional Neural Network Architectures0
WrapNet: Neural Net Inference with Ultra-Low-Resolution Arithmetic0
End-to-end Learning of Compressible Features0
A Post-coder Feedback Approach to Overcome Training Asymmetry in MIMO-TDD0
Analysis and Optimization for RIS-Aided Multi-Pair Communications Relying on Statistical CSI0
Byzantine-Resilient Secure Federated Learning0
Search What You Want: Barrier Panelty NAS for Mixed Precision Quantization0
Differentiable Joint Pruning and Quantization for Hardware Efficiency0
DBQ: A Differentiable Branch Quantizer for Lightweight Deep Neural Networks0
Exploiting vulnerabilities of deep neural networks for privacy protectionCode0
Training with reduced precision of a support vector machine model for text classification0
FSpiNN: An Optimization Framework for Memory- and Energy-Efficient Spiking Neural Networks0
eSampling: Energy Harvesting ADCs0
Compression strategies and space-conscious representations for deep neural networks0
A General Family of Stochastic Proximal Gradient Methods for Deep Learning0
Finding Non-Uniform Quantization Schemes using Multi-Task Gaussian ProcessesCode0
Image De-Quantization Using Generative Models as Priors0
Adaptive Periodic Averaging: A Practical Approach to Reducing Communication in Distributed Learning0
Experimental results on palmvein-based personal recognition by multi-snapshot fusion of textural features0
Term Revealing: Furthering Quantization at Run Time on Quantized DNNs0
Quantization in Relative Gradient Angle Domain For Building Polygon Estimation0
SGQuant: Squeezing the Last Bit on Graph Neural Networks with Specialized Quantization0
AUSN: Approximately Uniform Quantization by Adaptively Superimposing Non-uniform Distribution for Deep Neural Networks0
Enabling On-Device CNN Training by Self-Supervised Instance Filtering and Error Map Pruning0
Learning with tree tensor networks: complexity estimates and model selection0
Hardware Acceleration of Sparse and Irregular Tensor Computations of ML Models: A Survey and Insights0
Compressing Neural Machine Translation Models with 4-bit Precision0
Edinburgh's Submissions to the 2020 Machine Translation Efficiency Task0
Learning to Quantize Deep Neural Networks: A Competitive-Collaborative Approach0
Robust Product Markovian Quantization0
Show:102550
← PrevPage 78 of 99Next →

Benchmark Results

#ModelMetricClaimedVerifiedStatus
1FQ-ViT (ViT-L)Top-1 Accuracy (%)85.03Unverified
2FQ-ViT (ViT-B)Top-1 Accuracy (%)83.31Unverified
3FQ-ViT (Swin-B)Top-1 Accuracy (%)82.97Unverified
4FQ-ViT (Swin-S)Top-1 Accuracy (%)82.71Unverified
5FQ-ViT (DeiT-B)Top-1 Accuracy (%)81.2Unverified
6FQ-ViT (Swin-T)Top-1 Accuracy (%)80.51Unverified
7FQ-ViT (DeiT-S)Top-1 Accuracy (%)79.17Unverified
8Xception W8A8Top-1 Accuracy (%)78.97Unverified
9ADLIK-MO-ResNet50-W4A4Top-1 Accuracy (%)77.88Unverified
10ADLIK-MO-ResNet50-W3A4Top-1 Accuracy (%)77.34Unverified
#ModelMetricClaimedVerifiedStatus
13DCNN_VIVA_3MAP160,327.04Unverified
2DTQMAP0.79Unverified
#ModelMetricClaimedVerifiedStatus
1OutEffHop-Bert_basePerplexity6.3Unverified
2OutEffHop-Bert_basePerplexity6.21Unverified
#ModelMetricClaimedVerifiedStatus
1Accuracy98.13Unverified
#ModelMetricClaimedVerifiedStatus
1Accuracy92.92Unverified
#ModelMetricClaimedVerifiedStatus
1SSD ResNet50 V1 FPN 640x640MAP34.3Unverified
#ModelMetricClaimedVerifiedStatus
1TAR @ FAR=1e-495.13Unverified
#ModelMetricClaimedVerifiedStatus
1TAR @ FAR=1e-496.38Unverified
#ModelMetricClaimedVerifiedStatus
13DCNN_VIVA_5All84,809,664Unverified
#ModelMetricClaimedVerifiedStatus
1Accuracy99.8Unverified