SOTAVerified

Quantization

Quantization is a promising technique to reduce the computation cost of neural network training, which can replace high-cost floating-point numbers (e.g., float32) with low-cost fixed-point numbers (e.g., int8/int16).

Source: Adaptive Precision Training: Quantify Back Propagation in Neural Networks with Fixed-point Numbers

Papers

Showing 17011750 of 4925 papers

TitleStatusHype
Dynamic Signal Measurements Based on Quantized Data0
Communication-Efficient Federated Distillation0
Fast Autoregressive Models for Continuous Latent Generation0
Fast binary embeddings, and quantized compressed sensing with structured matrices0
Dynamic quantized consensus under DoS attacks: Towards a tight zooming-out factor0
Fast, Compact, and High Quality LSTM-RNN Based Statistical Parametric Speech Synthesizers for Mobile Devices0
Dynamic Quantized Consensus of General Linear Multi-agent Systems under Denial-of-Service Attacks0
Blind-Adaptive Quantizers0
Fast DistilBERT on CPUs0
FAST: DNN Training Under Variable Precision Block Floating Point with Stochastic Rounding0
Fastening the Initial Access in 5G NR Sidelink for 6G V2X Networks0
Arabic Compact Language Modelling for Resource Limited Devices0
Faster Inference of Integer SWIN Transformer by Removing the GELU Activation0
Faster Neural Net Inference via Forests of Sparse Oblique Decision Trees0
Communication-efficient k-Means for Edge-based Machine Learning0
FastICARL: Fast Incremental Classifier and Representation Learning with Efficient Budget Allocation in Audio Sensing Applications0
Fast Implementation of 4-bit Convolutional Neural Networks for Mobile Devices0
An Extra RMSNorm is All You Need for Fine Tuning to 1.58 Bits0
Dynamic Q&A of Clinical Documents with Large Language Models0
Fast Jet Tagging with MLP-Mixers on FPGAs0
Fast Large-Scale Discrete Optimization Based on Principal Coordinate Descent0
Fast learning rates with heavy-tailed losses0
Dynamic Probabilistic Pruning: A general framework for hardware-constrained pruning at different granularities0
Blending Low and High-Level Semantics of Time Series for Better Masked Time Series Generation0
FastMamba: A High-Speed and Efficient Mamba Accelerator on FPGA with Accurate Quantization0
Dynamic Predictive Sampling Analog to Digital Converter for Sparse Signal Sensing0
Blended Coarse Gradient Descent for Full Quantization of Deep Neural Networks0
ADaPTION: Toolbox and Benchmark for Training Convolutional Neural Networks with Reduced Numerical Precision Weights and Activation0
FQ-Conv: Fully Quantized Convolution for Efficient and Accurate Inference0
Adaptive Error-Bounded Hierarchical Matrices for Efficient Neural Network Compression0
Bi-ViT: Pushing the Limit of Vision Transformer Quantization0
FastSGD: A Fast Compressed SGD Framework for Distributed Machine Learning0
Dynamic Cell Modeling of Li-Ion Polymer Batteries for Precise SOC Estimation in Power-Needy Autonomous Electric Vehicles0
DyBit: Dynamic Bit-Precision Numbers for Efficient Quantized Neural Network Inference0
An Experimental Study of Reduced-Voltage Operation in Modern FPGAs for Neural Network Acceleration0
DV-Det: Efficient 3D Point Cloud Object Detection with Dynamic Voxelization0
Fast top-K Cosine Similarity Search through XOR-Friendly Binary Quantization on GPUs0
FAT: An In-Memory Accelerator with Fast Addition for Ternary Weight Neural Networks0
DupNet: Towards Very Tiny Quantized CNN with Improved Accuracy for Face Detection0
FATNN: Fast and Accurate Ternary Neural Networks0
Fault-Tolerant Four-Dimensional Constellation for Coherent Optical Transmission Systems0
BitTTS: Highly Compact Text-to-Speech Using 1.58-bit Quantization and Weight Indexing0
FBI: Fingerprinting models with Benign Inputs0
Compact recurrent neural networks for acoustic event detection on low-energy low-complexity platforms0
FBQuant: FeedBack Quantization for Large Language Models0
FCN-Pose: A Pruned and Quantized CNN for Robot Pose Estimation for Constrained Devices0
An Experimental Study: Assessing the Combined Framework of WavLM and BEST-RQ for Text-to-Speech Synthesis0
FD Cell-Free mMIMO: Analysis and Optimization0
Dual Precision Quantization for Efficient and Accurate Deep Neural Networks Inference0
An Exact Quantized Decentralized Gradient Descent Algorithm0
Show:102550
← PrevPage 35 of 99Next →

Benchmark Results

#ModelMetricClaimedVerifiedStatus
1FQ-ViT (ViT-L)Top-1 Accuracy (%)85.03Unverified
2FQ-ViT (ViT-B)Top-1 Accuracy (%)83.31Unverified
3FQ-ViT (Swin-B)Top-1 Accuracy (%)82.97Unverified
4FQ-ViT (Swin-S)Top-1 Accuracy (%)82.71Unverified
5FQ-ViT (DeiT-B)Top-1 Accuracy (%)81.2Unverified
6FQ-ViT (Swin-T)Top-1 Accuracy (%)80.51Unverified
7FQ-ViT (DeiT-S)Top-1 Accuracy (%)79.17Unverified
8Xception W8A8Top-1 Accuracy (%)78.97Unverified
9ADLIK-MO-ResNet50-W4A4Top-1 Accuracy (%)77.88Unverified
10ADLIK-MO-ResNet50-W3A4Top-1 Accuracy (%)77.34Unverified
#ModelMetricClaimedVerifiedStatus
13DCNN_VIVA_3MAP160,327.04Unverified
2DTQMAP0.79Unverified
#ModelMetricClaimedVerifiedStatus
1OutEffHop-Bert_basePerplexity6.3Unverified
2OutEffHop-Bert_basePerplexity6.21Unverified
#ModelMetricClaimedVerifiedStatus
1Accuracy98.13Unverified
#ModelMetricClaimedVerifiedStatus
1Accuracy92.92Unverified
#ModelMetricClaimedVerifiedStatus
1SSD ResNet50 V1 FPN 640x640MAP34.3Unverified
#ModelMetricClaimedVerifiedStatus
1TAR @ FAR=1e-495.13Unverified
#ModelMetricClaimedVerifiedStatus
1TAR @ FAR=1e-496.38Unverified
#ModelMetricClaimedVerifiedStatus
13DCNN_VIVA_5All84,809,664Unverified
#ModelMetricClaimedVerifiedStatus
1Accuracy99.8Unverified