SOTAVerified

Quantization

Quantization is a promising technique to reduce the computation cost of neural network training, which can replace high-cost floating-point numbers (e.g., float32) with low-cost fixed-point numbers (e.g., int8/int16).

Source: Adaptive Precision Training: Quantify Back Propagation in Neural Networks with Fixed-point Numbers

Papers

Showing 451500 of 4925 papers

TitleStatusHype
Context-aware Communication for Multi-agent Reinforcement LearningCode1
Mini-GPTs: Efficient Large Language Models through Contextual PruningCode1
Does Vector Quantization Fail in Spatio-Temporal Forecasting? Exploring a Differentiable Sparse Soft-Vector Quantization ApproachCode1
SmoothQuant+: Accurate and Efficient 4-bit Post-Training WeightQuantization for LLMCode1
Language Embedded 3D Gaussians for Open-Vocabulary Scene UnderstandingCode1
BAND-2k: Banding Artifact Noticeable Database for Banding Detection and Quality AssessmentCode1
TFMQ-DM: Temporal Feature Maintenance Quantization for Diffusion ModelsCode1
BinaryHPE: 3D Human Pose and Shape Estimation via BinarizationCode1
ComPEFT: Compression for Communicating Parameter Efficient Updates via Sparsification and QuantizationCode1
LQ-LoRA: Low-rank Plus Quantized Matrix Decomposition for Efficient Language Model FinetuningCode1
Generative De-Quantization for Neural Speech Codec via Latent DiffusionCode1
SynA-ResNet: Spike-driven ResNet Achieved through OR Residual ConnectionCode1
Quantized Distillation: Optimizing Driver Activity Recognition Models for Resource-Constrained EnvironmentsCode1
AFPQ: Asymmetric Floating Point Quantization for LLMsCode1
Embedding in Recommender Systems: A SurveyCode1
Edge AI-Based Vein Detector for Efficient Venipuncture in the Antecubital FossaCode1
Codebook Features: Sparse and Discrete Interpretability for Neural NetworksCode1
Structured Multi-Track Accompaniment Arrangement via Style Prior ModellingCode1
Hierarchical Vector Quantized Transformer for Multi-class Unsupervised Anomaly DetectionCode1
Matrix Compression via Randomized Low Rank and Low Precision FactorizationCode1
Watermarking LLMs with Weight QuantizationCode1
RoomDesigner: Encoding Anchor-latents for Style-consistent and Shape-compatible Indoor Scene GenerationCode1
QUIK: Towards End-to-End 4-Bit Inference on Generative Large Language ModelsCode1
QLLM: Accurate and Efficient Low-Bitwidth Quantization for Large Language ModelsCode1
Enhancing Text-based Knowledge Graph Completion with Zero-Shot Large Language Models: A Focus on Semantic EnhancementCode1
Sparse Fine-tuning for Inference Acceleration of Large Language ModelsCode1
EfficientDM: Efficient Quantization-Aware Fine-Tuning of Low-Bit Diffusion ModelsCode1
Compressing LLMs: The Truth is Rarely Pure and Never SimpleCode1
PB-LLM: Partially Binarized Large Language ModelsCode1
Pruning Small Pre-Trained Weights Irreversibly and Monotonically Impairs "Difficult" Downstream Tasks in LLMsCode1
RECOMBINER: Robust and Enhanced Compression with Bayesian Implicit Neural RepresentationsCode1
QDFormer: Towards Robust Audiovisual Segmentation in Complex Environments with Quantization-based Semantic DecompositionCode1
Revisiting Cephalometric Landmark Detection from the view of Human Pose Estimation with Lightweight Super-Resolution HeadCode1
Finite Scalar Quantization: VQ-VAE Made SimpleCode1
AsymFormer: Asymmetrical Cross-Modal Representation Learning for Mobile Platform Real-Time RGB-D Semantic SegmentationCode1
Model-enhanced Vector IndexCode1
CPLLM: Clinical Prediction with Large Language ModelsCode1
TransMUSIC: A Transformer-Aided Subspace Method for DOA Estimation with Low-Resolution ADCsCode1
Differentiable JPEG: The Devil is in the DetailsCode1
RawHash2: Mapping Raw Nanopore Signals Using Hash-Based Seeding and Adaptive QuantizationCode1
Channel Estimation for Quantized Systems based on Conditionally Gaussian Latent ModelsCode1
Few shot font generation via transferring similarity guided global style and quantization local styleCode1
RepCodec: A Speech Representation Codec for Speech TokenizationCode1
VQ-Font: Few-Shot Font Generation with Structure-Aware Enhancement and QuantizationCode1
Exploring Parameter-Efficient Fine-Tuning Techniques for Code Generation with Large Language ModelsCode1
Jumping through Local Minima: Quantization in the Loss Landscape of Vision TransformersCode1
NAPA-VQ: Neighborhood Aware Prototype Augmentation with Vector Quantization for Continual LearningCode1
EQ-Net: Elastic Quantization Neural NetworksCode1
Token-Scaled Logit Distillation for Ternary Weight Generative Language ModelsCode1
RMP-Loss: Regularizing Membrane Potential Distribution for Spiking Neural NetworksCode1
Show:102550
← PrevPage 10 of 99Next →

Benchmark Results

#ModelMetricClaimedVerifiedStatus
1FQ-ViT (ViT-L)Top-1 Accuracy (%)85.03Unverified
2FQ-ViT (ViT-B)Top-1 Accuracy (%)83.31Unverified
3FQ-ViT (Swin-B)Top-1 Accuracy (%)82.97Unverified
4FQ-ViT (Swin-S)Top-1 Accuracy (%)82.71Unverified
5FQ-ViT (DeiT-B)Top-1 Accuracy (%)81.2Unverified
6FQ-ViT (Swin-T)Top-1 Accuracy (%)80.51Unverified
7FQ-ViT (DeiT-S)Top-1 Accuracy (%)79.17Unverified
8Xception W8A8Top-1 Accuracy (%)78.97Unverified
9ADLIK-MO-ResNet50-W4A4Top-1 Accuracy (%)77.88Unverified
10ADLIK-MO-ResNet50-W3A4Top-1 Accuracy (%)77.34Unverified
#ModelMetricClaimedVerifiedStatus
13DCNN_VIVA_3MAP160,327.04Unverified
2DTQMAP0.79Unverified
#ModelMetricClaimedVerifiedStatus
1OutEffHop-Bert_basePerplexity6.3Unverified
2OutEffHop-Bert_basePerplexity6.21Unverified
#ModelMetricClaimedVerifiedStatus
1Accuracy98.13Unverified
#ModelMetricClaimedVerifiedStatus
1Accuracy92.92Unverified
#ModelMetricClaimedVerifiedStatus
1SSD ResNet50 V1 FPN 640x640MAP34.3Unverified
#ModelMetricClaimedVerifiedStatus
1TAR @ FAR=1e-495.13Unverified
#ModelMetricClaimedVerifiedStatus
1TAR @ FAR=1e-496.38Unverified
#ModelMetricClaimedVerifiedStatus
13DCNN_VIVA_5All84,809,664Unverified
#ModelMetricClaimedVerifiedStatus
1Accuracy99.8Unverified