SOTAVerified

Knowledge Distillation

Knowledge distillation is the process of transferring knowledge from a large model to a smaller one. While large models (such as very deep neural networks or ensembles of many models) have higher knowledge capacity than small models, this capacity might not be fully utilized.

Papers

Showing 23512375 of 4240 papers

TitleStatusHype
FEED: Feature-level Ensemble for Knowledge Distillation0
Few-shot 3D LiDAR Semantic Segmentation for Autonomous Driving0
Few-shot Face Image Translation via GAN Prior Distillation0
Few-shot learning of neural networks from scratch by pseudo example optimization0
Few-Shot Object Detection by Knowledge Distillation Using Bag-of-Visual-Words Representations0
Optimizing Vision Transformers with Data-Free Knowledge Transfer0
Optimizing YOLOv5s Object Detection through Knowledge Distillation algorithm0
Oracle Teacher: Leveraging Target Information for Better Knowledge Distillation of CTC Models0
Orderly Dual-Teacher Knowledge Distillation for Lightweight Human Pose Estimation0
ORQA: A Benchmark and Foundation Model for Holistic Operating Room Modeling0
Overcoming Language Priors for Visual Question Answering Based on Knowledge Distillation0
P4: Towards private, personalized, and Peer-to-Peer learning0
Pacemaker: Intermediate Teacher Knowledge Distillation For On-The-Fly Convolutional Neural Network0
PAIR: Leveraging Passage-Centric Similarity Relation for Improving Dense Passage Retrieval0
Pan-infection Foundation Framework Enables Multiple Pathogen Prediction0
PANLP at MEDIQA 2019: Pre-trained Language Models, Transfer Learning and Knowledge Distillation0
Papago’s Submission for the WMT21 Quality Estimation Shared Task0
Paralinguistic Privacy Protection at the Edge0
Parameter-Efficient and Student-Friendly Knowledge Distillation0
Parameter-Efficient Conformers via Sharing Sparsely-Gated Experts for End-to-End Speech Recognition0
Parameter Efficient Diverse Paraphrase Generation Using Sequence-Level Knowledge Distillation0
Partial Knowledge Distillation for Alleviating the Inherent Inter-Class Discrepancy in Federated Learning0
Partial to Whole Knowledge Distillation: Progressive Distilling Decomposed Knowledge Boosts Student Better0
PC-LoRA: Low-Rank Adaptation for Progressive Model Compression with Knowledge Distillation0
PDALN: Progressive Domain Adaptation over a Pre-trained Model for Low-Resource Cross-Domain Named Entity Recognition0
Show:102550
← PrevPage 95 of 170Next →

Benchmark Results

#ModelMetricClaimedVerifiedStatus
1ScaleKD (T:BEiT-L S:ViT-B/14)Top-1 accuracy %86.43Unverified
2ScaleKD (T:Swin-L S:ViT-B/16)Top-1 accuracy %85.53Unverified
3ScaleKD (T:Swin-L S:ViT-S/16)Top-1 accuracy %83.93Unverified
4ScaleKD (T:Swin-L S:Swin-T)Top-1 accuracy %83.8Unverified
5KD++(T: regnety-16GF S:ViT-B)Top-1 accuracy %83.6Unverified
6VkD (T:RegNety 160 S:DeiT-S)Top-1 accuracy %82.9Unverified
7SpectralKD (T:Swin-S S:Swin-T)Top-1 accuracy %82.7Unverified
8ScaleKD (T:Swin-L S:ResNet-50)Top-1 accuracy %82.55Unverified
9DiffKD (T:Swin-L S: Swin-T)Top-1 accuracy %82.5Unverified
10DIST (T: Swin-L S: Swin-T)Top-1 accuracy %82.3Unverified
#ModelMetricClaimedVerifiedStatus
1SRD (T:resnet-32x4, S:shufflenet-v2)Top-1 Accuracy (%)79.86Unverified
2shufflenet-v2(T:resnet-32x4, S:shufflenet-v2)Top-1 Accuracy (%)78.76Unverified
3MV-MR (T: CLIP/ViT-B-16 S: resnet50)Top-1 Accuracy (%)78.6Unverified
4resnet8x4 (T: resnet32x4 S: resnet8x4)Top-1 Accuracy (%)78.28Unverified
5resnet8x4 (T: resnet32x4 S: resnet8x4 [modified])Top-1 Accuracy (%)78.08Unverified
6ReviewKD++(T:resnet-32x4, S:shufflenet-v2)Top-1 Accuracy (%)77.93Unverified
7ReviewKD++(T:resnet-32x4, S:shufflenet-v1)Top-1 Accuracy (%)77.68Unverified
8resnet8x4 (T: resnet32x4 S: resnet8x4)Top-1 Accuracy (%)77.5Unverified
9resnet8x4 (T: resnet32x4 S: resnet8x4)Top-1 Accuracy (%)76.68Unverified
10resnet8x4 (T: resnet32x4 S: resnet8x4)Top-1 Accuracy (%)76.31Unverified
#ModelMetricClaimedVerifiedStatus
1LSHFM (T: ResNet101 S: ResNet50)mAP93.17Unverified
2LSHFM (T: ResNet101 S: MobileNetV2)mAP90.14Unverified
#ModelMetricClaimedVerifiedStatus
1TIE-KD (T: Adabins S: MobileNetV2)RMSE2.43Unverified