SOTAVerified

Knowledge Distillation

Knowledge distillation is the process of transferring knowledge from a large model to a smaller one. While large models (such as very deep neural networks or ensembles of many models) have higher knowledge capacity than small models, this capacity might not be fully utilized.

Papers

Showing 39764000 of 4240 papers

TitleStatusHype
ROSAR: An Adversarial Re-Training Framework for Robust Side-Scan Sonar Object DetectionCode0
Adversarial Moment-Matching Distillation of Large Language ModelsCode0
Delta Distillation for Efficient Video ProcessingCode0
AttriPrompter: Auto-Prompting with Attribute Semantics for Zero-shot Nuclei Detection via Visual-Language Pre-trained ModelsCode0
Few Shot Network Compression via Cross DistillationCode0
Sub-goal Distillation: A Method to Improve Small Language AgentsCode0
Few-shot Class-Incremental Semantic Segmentation via Pseudo-Labeling and Knowledge DistillationCode0
RUIE: Retrieval-based Unified Information Extraction using Large Language ModelCode0
deepQuest-py: Large and Distilled Models for Quality EstimationCode0
WAVER: Writing-style Agnostic Text-Video Retrieval via Distilling Vision-Language Models Through Open-Vocabulary KnowledgeCode0
Subspace Distillation for Continual LearningCode0
A Forward and Backward Compatible Framework for Few-shot Class-incremental Pill RecognitionCode0
"No Matter What You Do": Purifying GNN Models via Backdoor UnlearningCode0
Integrating Translation Memories into Non-Autoregressive Machine TranslationCode0
Non-Autoregressive Neural Machine TranslationCode0
Uncertainty-Guided Cross Attention Ensemble Mean Teacher for Semi-supervised Medical Image SegmentationCode0
CDFKD-MFS: Collaborative Data-free Knowledge Distillation via Multi-level Feature SharingCode0
Knowledge Distillation with Deep SupervisionCode0
Catch-Up Distillation: You Only Need to Train Once for Accelerating SamplingCode0
Few-Shot Class-Incremental Learning via Entropy-Regularized Data-Free ReplayCode0
FedSiKD: Clients Similarity and Knowledge Distillation: Addressing Non-i.i.d. and Constraints in Federated LearningCode0
Catastrophic Interference in Reinforcement Learning: A Solution Based on Context Division and Knowledge DistillationCode0
Exploring Vacant Classes in Label-Skewed Federated LearningCode0
SaiT: Sparse Vision Transformers through Adaptive Token PruningCode0
Not Far Away, Not So Close: Sample Efficient Nearest Neighbour Data Augmentation via MiniMaxCode0
Show:102550
← PrevPage 160 of 170Next →

Benchmark Results

#ModelMetricClaimedVerifiedStatus
1ScaleKD (T:BEiT-L S:ViT-B/14)Top-1 accuracy %86.43Unverified
2ScaleKD (T:Swin-L S:ViT-B/16)Top-1 accuracy %85.53Unverified
3ScaleKD (T:Swin-L S:ViT-S/16)Top-1 accuracy %83.93Unverified
4ScaleKD (T:Swin-L S:Swin-T)Top-1 accuracy %83.8Unverified
5KD++(T: regnety-16GF S:ViT-B)Top-1 accuracy %83.6Unverified
6VkD (T:RegNety 160 S:DeiT-S)Top-1 accuracy %82.9Unverified
7SpectralKD (T:Swin-S S:Swin-T)Top-1 accuracy %82.7Unverified
8ScaleKD (T:Swin-L S:ResNet-50)Top-1 accuracy %82.55Unverified
9DiffKD (T:Swin-L S: Swin-T)Top-1 accuracy %82.5Unverified
10DIST (T: Swin-L S: Swin-T)Top-1 accuracy %82.3Unverified
#ModelMetricClaimedVerifiedStatus
1SRD (T:resnet-32x4, S:shufflenet-v2)Top-1 Accuracy (%)79.86Unverified
2shufflenet-v2(T:resnet-32x4, S:shufflenet-v2)Top-1 Accuracy (%)78.76Unverified
3MV-MR (T: CLIP/ViT-B-16 S: resnet50)Top-1 Accuracy (%)78.6Unverified
4resnet8x4 (T: resnet32x4 S: resnet8x4)Top-1 Accuracy (%)78.28Unverified
5resnet8x4 (T: resnet32x4 S: resnet8x4 [modified])Top-1 Accuracy (%)78.08Unverified
6ReviewKD++(T:resnet-32x4, S:shufflenet-v2)Top-1 Accuracy (%)77.93Unverified
7ReviewKD++(T:resnet-32x4, S:shufflenet-v1)Top-1 Accuracy (%)77.68Unverified
8resnet8x4 (T: resnet32x4 S: resnet8x4)Top-1 Accuracy (%)77.5Unverified
9resnet8x4 (T: resnet32x4 S: resnet8x4)Top-1 Accuracy (%)76.68Unverified
10resnet8x4 (T: resnet32x4 S: resnet8x4)Top-1 Accuracy (%)76.31Unverified
#ModelMetricClaimedVerifiedStatus
1LSHFM (T: ResNet101 S: ResNet50)mAP93.17Unverified
2LSHFM (T: ResNet101 S: MobileNetV2)mAP90.14Unverified
#ModelMetricClaimedVerifiedStatus
1TIE-KD (T: Adabins S: MobileNetV2)RMSE2.43Unverified