SOTAVerified

Quantization

Quantization is a promising technique to reduce the computation cost of neural network training, which can replace high-cost floating-point numbers (e.g., float32) with low-cost fixed-point numbers (e.g., int8/int16).

Source: Adaptive Precision Training: Quantify Back Propagation in Neural Networks with Fixed-point Numbers

Papers

Showing 31013150 of 4925 papers

TitleStatusHype
Towards Semantic Communications: Deep Learning-Based Image Semantic Coding0
Towards Superior Quantization Accuracy: A Layer-sensitive Approach0
Towards the Limit of Network Quantization0
Towards Unified INT8 Training for Convolutional Neural Network0
Towards Unsupervised Speech Recognition and Synthesis with Quantized Speech Representation Learning0
Towards Variable and Coordinated Holistic Co-Speech Motion Generation0
Towards Watermarking of Open-Source LLMs0
TP-Aware Dequantization0
TQ-DiT: Efficient Time-Aware Quantization for Diffusion Transformers0
Trainable Fixed-Point Quantization for Deep Learning Acceleration on FPGAs0
Train Big, Then Compress: Rethinking Model Size for Efficient Training and Inference of Transformers0
Train Flat, Then Compress: Sharpness-Aware Minimization Learns More Compressible Models0
Training Acceleration of Low-Rank Decomposed Networks using Sequential Freezing and Rank Quantization0
Training and Inference for Integer-Based Semantic Segmentation Network0
Training DNN IoT Applications for Deployment On Analog NVM Crossbars0
Training Integer-Only Deep Recurrent Neural Networks0
Training of mixed-signal optical convolutional neural network with reduced quantization level0
Training Quantized Neural Networks with a Full-precision Auxiliary Module0
Training Quantized Neural Networks to Global Optimality via Semidefinite Programming0
Training with reduced precision of a support vector machine model for text classification0
Transceiver Cooperative Learning-aided Semantic Communications Against Mismatched Background Knowledge Bases0
Transferable Sequential Recommendation via Vector Quantized Meta Learning0
Transfer Hashing with Privileged Information0
Transformations in Learned Image Compression from a Modulation Perspective0
Transform-Based Feature Map Compression for CNN Inference0
Transformed Residual Quantization for Approximate Nearest Neighbor Search0
Transformer-based Clipped Contrastive Quantization Learning for Unsupervised Image Retrieval0
Transformer-based models and hardware acceleration analysis in autonomous driving: A survey0
Transformer-Based Nonlinear Transform Coding for Multi-Rate CSI Compression in MIMO-OFDM Systems0
Transformer-Lite: High-efficiency Deployment of Large Language Models on Mobile Phone GPUs0
Transform Quantization for CNN (Convolutional Neural Network) Compression0
Transition Rate Scheduling for Quantization-Aware Training0
Transmitting Data Through Reconfigurable Intelligent Surface: A Spatial Sigma-Delta Modulation Approach0
TR-DQ: Time-Rotation Diffusion Quantization0
T-RECX: Tiny-Resource Efficient Convolutional neural networks with early-eXit0
Tree Index: A New Cluster Evaluation Technique0
Tree Quantization for Large-Scale Similarity Search and Classification0
Triagem virtual de imagens de imuno-histoquímica usando redes neurais artificiais e espectro de padrões0
TrimLLM: Progressive Layer Dropping for Domain-Specific LLMs0
Trimming Down Large Spiking Vision Transformers via Heterogeneous Quantization Search0
TriplePlay: Enhancing Federated Learning with CLIP for Non-IID Data and Resource Efficiency0
TripleSpin - a generic compact paradigm for fast machine learning computations0
TROJAN-GUARD: Hardware Trojans Detection Using GNN in RTL Designs0
Truncated Non-Uniform Quantization for Distributed SGD0
Semantic-Enabled 6G Communication: A Task-oriented and Privacy-preserving Perspective0
TSPTQ-ViT: Two-scaled post-training quantization for vision transformer0
TTAQ: Towards Stable Post-training Quantization in Continuous Domain Adaptation0
TurboAttention: Efficient Attention Approximation For High Throughputs LLMs0
Turbo-ICL: In-Context Learning-Based Turbo Equalization0
TurboQuant: Online Vector Quantization with Near-optimal Distortion Rate0
Show:102550
← PrevPage 63 of 99Next →

Benchmark Results

#ModelMetricClaimedVerifiedStatus
1FQ-ViT (ViT-L)Top-1 Accuracy (%)85.03Unverified
2FQ-ViT (ViT-B)Top-1 Accuracy (%)83.31Unverified
3FQ-ViT (Swin-B)Top-1 Accuracy (%)82.97Unverified
4FQ-ViT (Swin-S)Top-1 Accuracy (%)82.71Unverified
5FQ-ViT (DeiT-B)Top-1 Accuracy (%)81.2Unverified
6FQ-ViT (Swin-T)Top-1 Accuracy (%)80.51Unverified
7FQ-ViT (DeiT-S)Top-1 Accuracy (%)79.17Unverified
8Xception W8A8Top-1 Accuracy (%)78.97Unverified
9ADLIK-MO-ResNet50-W4A4Top-1 Accuracy (%)77.88Unverified
10ADLIK-MO-ResNet50-W3A4Top-1 Accuracy (%)77.34Unverified
#ModelMetricClaimedVerifiedStatus
13DCNN_VIVA_3MAP160,327.04Unverified
2DTQMAP0.79Unverified
#ModelMetricClaimedVerifiedStatus
1OutEffHop-Bert_basePerplexity6.3Unverified
2OutEffHop-Bert_basePerplexity6.21Unverified
#ModelMetricClaimedVerifiedStatus
1Accuracy98.13Unverified
#ModelMetricClaimedVerifiedStatus
1Accuracy92.92Unverified
#ModelMetricClaimedVerifiedStatus
1SSD ResNet50 V1 FPN 640x640MAP34.3Unverified
#ModelMetricClaimedVerifiedStatus
1TAR @ FAR=1e-495.13Unverified
#ModelMetricClaimedVerifiedStatus
1TAR @ FAR=1e-496.38Unverified
#ModelMetricClaimedVerifiedStatus
13DCNN_VIVA_5All84,809,664Unverified
#ModelMetricClaimedVerifiedStatus
1Accuracy99.8Unverified