SOTAVerified

Quantization

Quantization is a promising technique to reduce the computation cost of neural network training, which can replace high-cost floating-point numbers (e.g., float32) with low-cost fixed-point numbers (e.g., int8/int16).

Source: Adaptive Precision Training: Quantify Back Propagation in Neural Networks with Fixed-point Numbers

Papers

Showing 38013850 of 4925 papers

TitleStatusHype
Bernoulli Embeddings for Graphs0
BeST -- A Novel Source Selection Metric for Transfer Learning0
Better Schedules for Low Precision Training of Deep Neural Networks0
Beyond Discreteness: Finite-Sample Analysis of Straight-Through Estimator for Quantization0
Beyond Neighbourhood-Preserving Transformations for Quantization-Based Unsupervised Hashing0
Beyond Quantization: Power aware neural networks0
Beyond Task Vectors: Selective Task Arithmetic Based on Importance Metrics0
Beyond the Tip of Efficiency: Uncovering the Submerged Threats of Jailbreak Attacks in Small Language Models0
Beyond Throughput and Compression Ratios: Towards High End-to-end Utility of Gradient Compression0
BF-IMNA: A Bit Fluid In-Memory Neural Architecture for Neural Network Acceleration0
BICM-compatible Rate Adaptive Geometric Constellation Shaping Using Optimized Many-to-one Labeling0
Bielik 11B v2 Technical Report0
Bifocal Neural ASR: Exploiting Keyword Spotting for Inference Optimization0
SpeedLimit: Neural Architecture Search for Quantized Transformer Models0
BiLiMO: Bit-Limited MIMO Radar via Task-Based Quantization0
Bilinear Random Projections for Locality-Sensitive Binary Codes0
Binarized Neural Network for Single Image Super Resolution0
Binarizing Sparse Convolutional Networks for Efficient Point Cloud Analysis0
BinaryBERT: Pushing the Limit of BERT Quantization0
Binary Constrained Deep Hashing Network for Image Retrieval without Manual Annotation0
Binary Neural Network for Speaker Verification0
Binary Neural Networks as a general-propose compute paradigm for on-device computer vision0
BinaryViT: Towards Efficient and Accurate Binary Vision Transformers0
Bioinspired Cortex-based Fast Codebook Generation0
Biologically Plausible Learning on Neuromorphic Hardware Architectures0
BiQGEMM: Matrix Multiplication with Lookup Table For Binary-Coding-based Quantized DNNs0
BiSup: Bidirectional Quantization Error Suppression for Large Language Models0
BiTAT: Neural Network Binarization with Task-dependent Aggregated Transformation0
Bit Efficient Quantization for Deep Neural Networks0
Bit-Mixer: Mixed-precision networks with runtime bit-width selection0
BitNet b1.58 Reloaded: State-of-the-art Performance Also on Smaller Networks0
BitPruning: Learning Bitlengths for Aggressive and Accurate Quantization0
BitsFusion: 1.99 bits Weight Quantization of Diffusion Model0
Bit-Shrinking: Limiting Instantaneous Sharpness for Improving Post-Training Quantization0
BitTTS: Highly Compact Text-to-Speech Using 1.58-bit Quantization and Weight Indexing0
Bi-ViT: Pushing the Limit of Vision Transformer Quantization0
Blended Coarse Gradient Descent for Full Quantization of Deep Neural Networks0
Blending Low and High-Level Semantics of Time Series for Better Masked Time Series Generation0
Blind-Adaptive Quantizers0
Block Modulating Video Compression: An Ultra Low Complexity Image Compression Encoder for Resource Limited Platforms0
Blockwise Compression of Transformer-based Models without Retraining0
Block-Wise Dynamic-Precision Neural Network Training Acceleration via Online Quantization Sensitivity Analytics0
BlueLM-V-3B: Algorithm and System Co-Design for Multimodal Large Language Models on Mobile Devices0
BMPQ: Bit-Gradient Sensitivity Driven Mixed-Precision Quantization of DNNs from Scratch0
BOMP-NAS: Bayesian Optimization Mixed Precision NAS0
Boost CTR Prediction for New Advertisements via Modeling Visual Content0
Boosted Dense Retriever0
Boosted Dense Retriever0
Boosting Distributed Full-graph GNN Training with Asynchronous One-bit Communication0
Boost Vision Transformer with GPU-Friendly Sparsity and Quantization0
Show:102550
← PrevPage 77 of 99Next →

Benchmark Results

#ModelMetricClaimedVerifiedStatus
1FQ-ViT (ViT-L)Top-1 Accuracy (%)85.03Unverified
2FQ-ViT (ViT-B)Top-1 Accuracy (%)83.31Unverified
3FQ-ViT (Swin-B)Top-1 Accuracy (%)82.97Unverified
4FQ-ViT (Swin-S)Top-1 Accuracy (%)82.71Unverified
5FQ-ViT (DeiT-B)Top-1 Accuracy (%)81.2Unverified
6FQ-ViT (Swin-T)Top-1 Accuracy (%)80.51Unverified
7FQ-ViT (DeiT-S)Top-1 Accuracy (%)79.17Unverified
8Xception W8A8Top-1 Accuracy (%)78.97Unverified
9ADLIK-MO-ResNet50-W4A4Top-1 Accuracy (%)77.88Unverified
10ADLIK-MO-ResNet50-W3A4Top-1 Accuracy (%)77.34Unverified
#ModelMetricClaimedVerifiedStatus
13DCNN_VIVA_3MAP160,327.04Unverified
2DTQMAP0.79Unverified
#ModelMetricClaimedVerifiedStatus
1OutEffHop-Bert_basePerplexity6.3Unverified
2OutEffHop-Bert_basePerplexity6.21Unverified
#ModelMetricClaimedVerifiedStatus
1Accuracy98.13Unverified
#ModelMetricClaimedVerifiedStatus
1Accuracy92.92Unverified
#ModelMetricClaimedVerifiedStatus
1SSD ResNet50 V1 FPN 640x640MAP34.3Unverified
#ModelMetricClaimedVerifiedStatus
1TAR @ FAR=1e-495.13Unverified
#ModelMetricClaimedVerifiedStatus
1TAR @ FAR=1e-496.38Unverified
#ModelMetricClaimedVerifiedStatus
13DCNN_VIVA_5All84,809,664Unverified
#ModelMetricClaimedVerifiedStatus
1Accuracy99.8Unverified