SOTAVerified

Knowledge Distillation

Knowledge distillation is the process of transferring knowledge from a large model to a smaller one. While large models (such as very deep neural networks or ensembles of many models) have higher knowledge capacity than small models, this capacity might not be fully utilized.

Papers

Showing 25762600 of 4240 papers

TitleStatusHype
A Comprehensive Overhaul of Distilling Unconditional GANs0
A Comprehensive Review of Knowledge Distillation in Computer Vision0
A Comprehensive Study of Class Incremental Learning Algorithms for Visual Tasks0
A Comprehensive Survey of Compression Algorithms for Language Models0
A Comprehensive Survey on Knowledge Distillation of Diffusion Models0
A Continual and Incremental Learning Approach for TinyML On-device Training Using Dataset Distillation and Model Size Adaption0
A Contrastive Teacher-Student Framework for Novelty Detection under Style Shifts0
Acquiring Knowledge from Pre-trained Model to Neural Machine Translation0
A Cross-Domain Approach for Continuous Impression Recognition from Dyadic Audio-Visual-Physio Signals0
Action Spotting and Precise Event Detection in Sports: Datasets, Methods, and Challenges0
Activation Map Adaptation for Effective Knowledge Distillation0
Active Class Incremental Learning for Imbalanced Datasets0
Active Data Curation Effectively Distills Large-Scale Multimodal Models0
Active Exploration of Multimodal Complementarity for Few-Shot Action Recognition0
Active Large Language Model-based Knowledge Distillation for Session-based Recommendation0
Active Learning for Lane Detection: A Knowledge Distillation Approach0
ActivityCLIP: Enhancing Group Activity Recognition by Mining Complementary Information from Text to Supplement Image Modality0
Ada-DQA: Adaptive Diverse Quality-aware Feature Acquisition for Video Quality Assessment0
AdaKD: Dynamic Knowledge Distillation of ASR models using Adaptive Loss Weighting0
Adam: Dense Retrieval Distillation with Adaptive Dark Examples0
Adapt-and-Distill: Developing Small, Fast and Effective Pretrained Language Models for Domains0
Adapter-based Selective Knowledge Distillation for Federated Multi-domain Meeting Summarization0
AdapterDistillation: Non-Destructive Task Composition with Knowledge Distillation0
Adapting Models to Signal Degradation using Distillation0
Adapting OC20-trained EquiformerV2 Models for High-Entropy Materials0
Show:102550
← PrevPage 104 of 170Next →

Benchmark Results

#ModelMetricClaimedVerifiedStatus
1ScaleKD (T:BEiT-L S:ViT-B/14)Top-1 accuracy %86.43Unverified
2ScaleKD (T:Swin-L S:ViT-B/16)Top-1 accuracy %85.53Unverified
3ScaleKD (T:Swin-L S:ViT-S/16)Top-1 accuracy %83.93Unverified
4ScaleKD (T:Swin-L S:Swin-T)Top-1 accuracy %83.8Unverified
5KD++(T: regnety-16GF S:ViT-B)Top-1 accuracy %83.6Unverified
6VkD (T:RegNety 160 S:DeiT-S)Top-1 accuracy %82.9Unverified
7SpectralKD (T:Swin-S S:Swin-T)Top-1 accuracy %82.7Unverified
8ScaleKD (T:Swin-L S:ResNet-50)Top-1 accuracy %82.55Unverified
9DiffKD (T:Swin-L S: Swin-T)Top-1 accuracy %82.5Unverified
10DIST (T: Swin-L S: Swin-T)Top-1 accuracy %82.3Unverified
#ModelMetricClaimedVerifiedStatus
1SRD (T:resnet-32x4, S:shufflenet-v2)Top-1 Accuracy (%)79.86Unverified
2shufflenet-v2(T:resnet-32x4, S:shufflenet-v2)Top-1 Accuracy (%)78.76Unverified
3MV-MR (T: CLIP/ViT-B-16 S: resnet50)Top-1 Accuracy (%)78.6Unverified
4resnet8x4 (T: resnet32x4 S: resnet8x4)Top-1 Accuracy (%)78.28Unverified
5resnet8x4 (T: resnet32x4 S: resnet8x4 [modified])Top-1 Accuracy (%)78.08Unverified
6ReviewKD++(T:resnet-32x4, S:shufflenet-v2)Top-1 Accuracy (%)77.93Unverified
7ReviewKD++(T:resnet-32x4, S:shufflenet-v1)Top-1 Accuracy (%)77.68Unverified
8resnet8x4 (T: resnet32x4 S: resnet8x4)Top-1 Accuracy (%)77.5Unverified
9resnet8x4 (T: resnet32x4 S: resnet8x4)Top-1 Accuracy (%)76.68Unverified
10resnet8x4 (T: resnet32x4 S: resnet8x4)Top-1 Accuracy (%)76.31Unverified
#ModelMetricClaimedVerifiedStatus
1LSHFM (T: ResNet101 S: ResNet50)mAP93.17Unverified
2LSHFM (T: ResNet101 S: MobileNetV2)mAP90.14Unverified
#ModelMetricClaimedVerifiedStatus
1TIE-KD (T: Adabins S: MobileNetV2)RMSE2.43Unverified