SOTAVerified

Knowledge Distillation

Knowledge distillation is the process of transferring knowledge from a large model to a smaller one. While large models (such as very deep neural networks or ensembles of many models) have higher knowledge capacity than small models, this capacity might not be fully utilized.

Papers

Showing 401425 of 4240 papers

TitleStatusHype
Backdoor Attacks on Self-Supervised LearningCode1
Backdoor Cleansing with Unlabeled DataCode1
DARTS: Double Attention Reference-based Transformer for Super-resolutionCode1
Faster ILOD: Incremental Learning for Object Detectors based on Faster RCNNCode1
Dark Experience for General Continual Learning: a Strong, Simple BaselineCode1
Fast Neural Architecture Search of Compact Semantic Segmentation Models via Auxiliary CellsCode1
Distilling the Knowledge of BERT for Sequence-to-Sequence ASRCode1
Balanced Knowledge Distillation for Long-tailed LearningCode1
Data Diversification: A Simple Strategy For Neural Machine TranslationCode1
DASS: Distilled Audio State Space Models Are Stronger and More Duration-Scalable LearnersCode1
f-Divergence Minimization for Sequence-Level Knowledge DistillationCode1
Feature Structure Distillation with Centered Kernel Alignment in BERT TransferringCode1
Aligned Structured Sparsity Learning for Efficient Image Super-ResolutionCode1
Data-Free Class-Incremental Hand Gesture RecognitionCode1
CaMEL: Mean Teacher Learning for Image CaptioningCode1
Data-Free Network Quantization With Adversarial Knowledge DistillationCode1
Data-Free Knowledge Distillation via Feature Exchange and Activation Region ConstraintCode1
BearingPGA-Net: A Lightweight and Deployable Bearing Fault Diagnosis Network via Decoupled Knowledge Distillation and FPGA AccelerationCode1
DA-Mamba: Domain Adaptive Hybrid Mamba-Transformer Based One-Stage Object DetectionCode1
FedMD: Heterogenous Federated Learning via Model DistillationCode1
FedSOL: Stabilized Orthogonal Learning with Proximal Restrictions in Federated LearningCode1
FedUKD: Federated UNet Model with Knowledge Distillation for Land Use Classification from Satellite and Street ViewsCode1
Align-KD: Distilling Cross-Modal Alignment Knowledge for Mobile Vision-Language ModelCode1
Fine-tuning Global Model via Data-Free Knowledge Distillation for Non-IID Federated LearningCode1
3D Annotation-Free Learning by Distilling 2D Open-Vocabulary Segmentation Models for Autonomous DrivingCode1
Show:102550
← PrevPage 17 of 170Next →

Benchmark Results

#ModelMetricClaimedVerifiedStatus
1ScaleKD (T:BEiT-L S:ViT-B/14)Top-1 accuracy %86.43Unverified
2ScaleKD (T:Swin-L S:ViT-B/16)Top-1 accuracy %85.53Unverified
3ScaleKD (T:Swin-L S:ViT-S/16)Top-1 accuracy %83.93Unverified
4ScaleKD (T:Swin-L S:Swin-T)Top-1 accuracy %83.8Unverified
5KD++(T: regnety-16GF S:ViT-B)Top-1 accuracy %83.6Unverified
6VkD (T:RegNety 160 S:DeiT-S)Top-1 accuracy %82.9Unverified
7SpectralKD (T:Swin-S S:Swin-T)Top-1 accuracy %82.7Unverified
8ScaleKD (T:Swin-L S:ResNet-50)Top-1 accuracy %82.55Unverified
9DiffKD (T:Swin-L S: Swin-T)Top-1 accuracy %82.5Unverified
10DIST (T: Swin-L S: Swin-T)Top-1 accuracy %82.3Unverified
#ModelMetricClaimedVerifiedStatus
1SRD (T:resnet-32x4, S:shufflenet-v2)Top-1 Accuracy (%)79.86Unverified
2shufflenet-v2(T:resnet-32x4, S:shufflenet-v2)Top-1 Accuracy (%)78.76Unverified
3MV-MR (T: CLIP/ViT-B-16 S: resnet50)Top-1 Accuracy (%)78.6Unverified
4resnet8x4 (T: resnet32x4 S: resnet8x4)Top-1 Accuracy (%)78.28Unverified
5resnet8x4 (T: resnet32x4 S: resnet8x4 [modified])Top-1 Accuracy (%)78.08Unverified
6ReviewKD++(T:resnet-32x4, S:shufflenet-v2)Top-1 Accuracy (%)77.93Unverified
7ReviewKD++(T:resnet-32x4, S:shufflenet-v1)Top-1 Accuracy (%)77.68Unverified
8resnet8x4 (T: resnet32x4 S: resnet8x4)Top-1 Accuracy (%)77.5Unverified
9resnet8x4 (T: resnet32x4 S: resnet8x4)Top-1 Accuracy (%)76.68Unverified
10resnet8x4 (T: resnet32x4 S: resnet8x4)Top-1 Accuracy (%)76.31Unverified
#ModelMetricClaimedVerifiedStatus
1LSHFM (T: ResNet101 S: ResNet50)mAP93.17Unverified
2LSHFM (T: ResNet101 S: MobileNetV2)mAP90.14Unverified
#ModelMetricClaimedVerifiedStatus
1TIE-KD (T: Adabins S: MobileNetV2)RMSE2.43Unverified