SOTAVerified

Quantization

Quantization is a promising technique to reduce the computation cost of neural network training, which can replace high-cost floating-point numbers (e.g., float32) with low-cost fixed-point numbers (e.g., int8/int16).

Source: Adaptive Precision Training: Quantify Back Propagation in Neural Networks with Fixed-point Numbers

Papers

Showing 30513100 of 4925 papers

TitleStatusHype
Energy Efficient Federated Learning over Heterogeneous Mobile Devices via Joint Design of Weight Quantization and Wireless Transmission0
Toward a Diffusion-Based Generalist for Dense Vision Tasks0
Toward Compact Parameter Representations for Architecture-Agnostic Neural Network Compression0
Toward Efficient Deep Spiking Neuron Networks:A Survey On Compression0
Toward Efficient Low-Precision Training: Data Format Optimization and Hysteresis Quantization0
Toward Extremely Low Bit and Lossless Accuracy in DNNs with Progressive ADMM0
Toward Fast and Accurate Neural Chinese Word Segmentation with Multi-Criteria Learning0
Toward INT4 Fixed-Point Training via Exploring Quantization Error for Gradients0
Toward Joint Image Generation and Compression using Generative Adversarial Networks0
Toward nonlinear dynamic control over encrypted data for infinite time horizon0
Towards Accurate and Efficient Sub-8-Bit Integer Training0
Towards Accurate and High-Speed Spiking Neuromorphic Systems with Data Quantization-Aware Deep Networks0
Towards Accurate Post-Training Quantization for Vision Transformer0
Towards Accurate Quantization and Pruning via Data-free Knowledge Transfer0
Towards AI-Native Fronthaul: Neural Compression for NextG Cloud RAN0
Towards a tailored mixed-precision sub-8-bit quantization scheme for Gated Recurrent Units using Genetic Algorithms0
Towards Cheaper Inference in Deep Networks with Lower Bit-Width Accumulators0
Towards Clip-Free Quantized Super-Resolution Networks: How to Tame Representative Images0
Towards Compact and Fast Neural Machine Translation Using a Combined Method0
Towards Effective 2-bit Quantization: Pareto-optimal Bit Allocation for Deep CNNs Compression0
Towards efficient deep autoencoders for multivariate time series anomaly detection0
Towards Efficient Full 8-bit Integer DNN Online Training on Resource-limited Devices without Batch Normalization0
Towards Efficient In-memory Computing Hardware for Quantized Neural Networks: State-of-the-art, Open Challenges and Perspectives0
Towards Efficient Post-training Quantization of Pre-trained Language Models0
Towards Efficient Pre-training: Exploring FP4 Precision in Large Language Models0
Towards Energy-Efficient and Secure Edge AI: A Cross-Layer Framework0
Towards Energy Efficient Federated Learning over 5G+ Mobile Devices0
HYPER-SNN: Towards Energy-efficient Quantized Deep Spiking Neural Networks for Hyperspectral Image Classification0
Towards Fast and Accurate Neural Chinese Word Segmentation with Multi-Criteria Learning0
Towards Fully 8-bit Integer Inference for the Transformer Model0
Towards Hardware-Specific Automatic Compression of Neural Networks0
Towards Improved Text-Aligned Codebook Learning: Multi-Hierarchical Codebook-Text Alignment with Long Text0
Towards Intelligent Millimeter and Terahertz Communication for 6G: Computer Vision-aided Beamforming0
Towards Learning Discrete Representations via Self-Supervision for Wearables-Based Human Activity Recognition0
Towards Lightweight Speaker Verification via Adaptive Neural Network Quantization0
Towards Low-bit Communication for Tensor Parallel LLM Inference0
Exploring the Potential of Low-bit Training of Convolutional Neural Networks0
Towards Low-loss 1-bit Quantization of User-item Representations for Top-K Recommendation0
Towards Mixed-Precision Quantization of Neural Networks via Constrained Optimization0
Model-Free Learning for the Linear Quadratic Regulator over Rate-Limited Channels0
Towards Neural Variational Monte Carlo That Scales Linearly with System Size0
Towards Next-Level Post-Training Quantization of Hyper-Scale Transformers0
Towards On-Device Face Recognition in Body-worn Cameras0
Towards Optimal Compression: Joint Pruning and Quantization0
Towards Practical and Efficient Image-to-Speech Captioning with Vision-Language Pre-training and Multi-modal Tokens0
Towards Practical Single-shot Phase Retrieval with Physics-Driven Deep Neural Network0
Towards Real-Time Neural Video Codec for Cross-Platform Application Using Calibration Information0
Towards Real-Time Neural Volumetric Rendering on Mobile Devices: A Measurement Study0
Towards Reasoning Ability of Small Language Models0
Towards Robust Low Light Image Enhancement0
Show:102550
← PrevPage 62 of 99Next →

Benchmark Results

#ModelMetricClaimedVerifiedStatus
1FQ-ViT (ViT-L)Top-1 Accuracy (%)85.03Unverified
2FQ-ViT (ViT-B)Top-1 Accuracy (%)83.31Unverified
3FQ-ViT (Swin-B)Top-1 Accuracy (%)82.97Unverified
4FQ-ViT (Swin-S)Top-1 Accuracy (%)82.71Unverified
5FQ-ViT (DeiT-B)Top-1 Accuracy (%)81.2Unverified
6FQ-ViT (Swin-T)Top-1 Accuracy (%)80.51Unverified
7FQ-ViT (DeiT-S)Top-1 Accuracy (%)79.17Unverified
8Xception W8A8Top-1 Accuracy (%)78.97Unverified
9ADLIK-MO-ResNet50-W4A4Top-1 Accuracy (%)77.88Unverified
10ADLIK-MO-ResNet50-W3A4Top-1 Accuracy (%)77.34Unverified
#ModelMetricClaimedVerifiedStatus
13DCNN_VIVA_3MAP160,327.04Unverified
2DTQMAP0.79Unverified
#ModelMetricClaimedVerifiedStatus
1OutEffHop-Bert_basePerplexity6.3Unverified
2OutEffHop-Bert_basePerplexity6.21Unverified
#ModelMetricClaimedVerifiedStatus
1Accuracy98.13Unverified
#ModelMetricClaimedVerifiedStatus
1Accuracy92.92Unverified
#ModelMetricClaimedVerifiedStatus
1SSD ResNet50 V1 FPN 640x640MAP34.3Unverified
#ModelMetricClaimedVerifiedStatus
1TAR @ FAR=1e-495.13Unverified
#ModelMetricClaimedVerifiedStatus
1TAR @ FAR=1e-496.38Unverified
#ModelMetricClaimedVerifiedStatus
13DCNN_VIVA_5All84,809,664Unverified
#ModelMetricClaimedVerifiedStatus
1Accuracy99.8Unverified