SOTAVerified

Knowledge Distillation

Knowledge distillation is the process of transferring knowledge from a large model to a smaller one. While large models (such as very deep neural networks or ensembles of many models) have higher knowledge capacity than small models, this capacity might not be fully utilized.

Papers

Showing 401425 of 4240 papers

TitleStatusHype
Backdoor Attacks on Self-Supervised LearningCode1
Backdoor Cleansing with Unlabeled DataCode1
Continual Learning for Image Segmentation with Dynamic QueryCode1
Contrastive Deep SupervisionCode1
EfficientBERT: Progressively Searching Multilayer Perceptron via Warm-up Knowledge DistillationCode1
Efficient Fine-Tuning and Concept Suppression for Pruned Diffusion ModelsCode1
Cross-Layer Distillation with Semantic CalibrationCode1
DASS: Distilled Audio State Space Models Are Stronger and More Duration-Scalable LearnersCode1
Efficient On-Device Session-Based RecommendationCode1
Content-Variant Reference Image Quality Assessment via Knowledge DistillationCode1
Learning Efficient Vision Transformers via Fine-Grained Manifold DistillationCode1
EfficientVLM: Fast and Accurate Vision-Language Models via Knowledge Distillation and Modal-adaptive PruningCode1
Aligned Structured Sparsity Learning for Efficient Image Super-ResolutionCode1
Eliminating Backdoor Triggers for Deep Neural Networks Using Attention Relation Graph DistillationCode1
Bayes Conditional Distribution Estimation for Knowledge Distillation Based on Conditional Mutual InformationCode1
Bridge Past and Future: Overcoming Information Asymmetry in Incremental Object DetectionCode1
Enhancing Low-resolution Face Recognition with Feature Similarity Knowledge DistillationCode1
BearingPGA-Net: A Lightweight and Deployable Bearing Fault Diagnosis Network via Decoupled Knowledge Distillation and FPGA AccelerationCode1
Content-Aware GAN CompressionCode1
Context-Aware Image Inpainting with Learned Semantic PriorsCode1
Are Intermediate Layers and Labels Really Necessary? A General Language Model Distillation MethodCode1
Designing Large Foundation Models for Efficient Training and Inference: A SurveyCode1
Align-KD: Distilling Cross-Modal Alignment Knowledge for Mobile Vision-Language ModelCode1
Exploring Complementary Strengths of Invariant and Equivariant Representations for Few-Shot LearningCode1
Continual All-in-One Adverse Weather Removal with Knowledge Replay on a Unified Network StructureCode1
Show:102550
← PrevPage 17 of 170Next →

Benchmark Results

#ModelMetricClaimedVerifiedStatus
1ScaleKD (T:BEiT-L S:ViT-B/14)Top-1 accuracy %86.43Unverified
2ScaleKD (T:Swin-L S:ViT-B/16)Top-1 accuracy %85.53Unverified
3ScaleKD (T:Swin-L S:ViT-S/16)Top-1 accuracy %83.93Unverified
4ScaleKD (T:Swin-L S:Swin-T)Top-1 accuracy %83.8Unverified
5KD++(T: regnety-16GF S:ViT-B)Top-1 accuracy %83.6Unverified
6VkD (T:RegNety 160 S:DeiT-S)Top-1 accuracy %82.9Unverified
7SpectralKD (T:Swin-S S:Swin-T)Top-1 accuracy %82.7Unverified
8ScaleKD (T:Swin-L S:ResNet-50)Top-1 accuracy %82.55Unverified
9DiffKD (T:Swin-L S: Swin-T)Top-1 accuracy %82.5Unverified
10DIST (T: Swin-L S: Swin-T)Top-1 accuracy %82.3Unverified
#ModelMetricClaimedVerifiedStatus
1SRD (T:resnet-32x4, S:shufflenet-v2)Top-1 Accuracy (%)79.86Unverified
2shufflenet-v2(T:resnet-32x4, S:shufflenet-v2)Top-1 Accuracy (%)78.76Unverified
3MV-MR (T: CLIP/ViT-B-16 S: resnet50)Top-1 Accuracy (%)78.6Unverified
4resnet8x4 (T: resnet32x4 S: resnet8x4)Top-1 Accuracy (%)78.28Unverified
5resnet8x4 (T: resnet32x4 S: resnet8x4 [modified])Top-1 Accuracy (%)78.08Unverified
6ReviewKD++(T:resnet-32x4, S:shufflenet-v2)Top-1 Accuracy (%)77.93Unverified
7ReviewKD++(T:resnet-32x4, S:shufflenet-v1)Top-1 Accuracy (%)77.68Unverified
8resnet8x4 (T: resnet32x4 S: resnet8x4)Top-1 Accuracy (%)77.5Unverified
9resnet8x4 (T: resnet32x4 S: resnet8x4)Top-1 Accuracy (%)76.68Unverified
10resnet8x4 (T: resnet32x4 S: resnet8x4)Top-1 Accuracy (%)76.31Unverified
#ModelMetricClaimedVerifiedStatus
1LSHFM (T: ResNet101 S: ResNet50)mAP93.17Unverified
2LSHFM (T: ResNet101 S: MobileNetV2)mAP90.14Unverified
#ModelMetricClaimedVerifiedStatus
1TIE-KD (T: Adabins S: MobileNetV2)RMSE2.43Unverified