SOTAVerified

Knowledge Distillation

Knowledge distillation is the process of transferring knowledge from a large model to a smaller one. While large models (such as very deep neural networks or ensembles of many models) have higher knowledge capacity than small models, this capacity might not be fully utilized.

Papers

Showing 14011425 of 4240 papers

TitleStatusHype
Multi-Branch Mutual-Distillation Transformer for EEG-Based Seizure Subtype Classification0
Distillation of Diffusion Features for Semantic Correspondence0
Mutli-View 3D Reconstruction using Knowledge DistillationCode0
QABISAR: Query-Article Bipartite Interactions for Statutory Article Retrieval0
Local vs. Global: Local Land-Use and Land-Cover Models Deliver Higher Quality Maps0
Continuous Concepts Removal in Text-to-image Diffusion Models0
Toward Fair Graph Neural Networks Via Dual-Teacher Knowledge Distillation0
Reverse Thinking Makes LLMs Stronger Reasoners0
Zero-shot Slot Filling in the Age of LLMs for Dialogue Systems0
Puzzle: Distillation-Based NAS for Inference-Optimized LLMs0
Headache to Overstock? Promoting Long-tail Items through Debiased Product Bundling0
Pre-Training Graph Contrastive Masked Autoencoders are Strong Distillers for EEG0
Improved implicit diffusion model with knowledge distillation to estimate the spatial distribution density of carbon stock in remote sensing imagery0
Active Data Curation Effectively Distills Large-Scale Multimodal Models0
Words Matter: Leveraging Individual Text Embeddings for Code Generation in CLIP Test-Time AdaptationCode0
Large-Scale Data-Free Knowledge Distillation for ImageNet via Multi-Resolution Data GenerationCode0
Beyond Task Vectors: Selective Task Arithmetic Based on Importance Metrics0
Ensemble Learning via Knowledge Transfer for CTR PredictionCode0
Dynamic Self-Distillation via Previous Mini-batches for Fine-tuning Small Language Models0
When Babies Teach Babies: Can student knowledge sharing outperform Teacher-Guided Distillation on small datasets?Code0
Leveraging Foundation Models To learn the shape of semi-fluid deformable objects0
TransFair: Transferring Fairness from Ocular Disease Classification to Progression Prediction0
Efficient Ternary Weight Embedding Model: Bridging Scalability and PerformanceCode0
Partial Knowledge Distillation for Alleviating the Inherent Inter-Class Discrepancy in Federated Learning0
Improving Mathematical Reasoning Capabilities of Small Language Models via Feedback-Driven Distillation0
Show:102550
← PrevPage 57 of 170Next →

Benchmark Results

#ModelMetricClaimedVerifiedStatus
1ScaleKD (T:BEiT-L S:ViT-B/14)Top-1 accuracy %86.43Unverified
2ScaleKD (T:Swin-L S:ViT-B/16)Top-1 accuracy %85.53Unverified
3ScaleKD (T:Swin-L S:ViT-S/16)Top-1 accuracy %83.93Unverified
4ScaleKD (T:Swin-L S:Swin-T)Top-1 accuracy %83.8Unverified
5KD++(T: regnety-16GF S:ViT-B)Top-1 accuracy %83.6Unverified
6VkD (T:RegNety 160 S:DeiT-S)Top-1 accuracy %82.9Unverified
7SpectralKD (T:Swin-S S:Swin-T)Top-1 accuracy %82.7Unverified
8ScaleKD (T:Swin-L S:ResNet-50)Top-1 accuracy %82.55Unverified
9DiffKD (T:Swin-L S: Swin-T)Top-1 accuracy %82.5Unverified
10DIST (T: Swin-L S: Swin-T)Top-1 accuracy %82.3Unverified
#ModelMetricClaimedVerifiedStatus
1SRD (T:resnet-32x4, S:shufflenet-v2)Top-1 Accuracy (%)79.86Unverified
2shufflenet-v2(T:resnet-32x4, S:shufflenet-v2)Top-1 Accuracy (%)78.76Unverified
3MV-MR (T: CLIP/ViT-B-16 S: resnet50)Top-1 Accuracy (%)78.6Unverified
4resnet8x4 (T: resnet32x4 S: resnet8x4)Top-1 Accuracy (%)78.28Unverified
5resnet8x4 (T: resnet32x4 S: resnet8x4 [modified])Top-1 Accuracy (%)78.08Unverified
6ReviewKD++(T:resnet-32x4, S:shufflenet-v2)Top-1 Accuracy (%)77.93Unverified
7ReviewKD++(T:resnet-32x4, S:shufflenet-v1)Top-1 Accuracy (%)77.68Unverified
8resnet8x4 (T: resnet32x4 S: resnet8x4)Top-1 Accuracy (%)77.5Unverified
9resnet8x4 (T: resnet32x4 S: resnet8x4)Top-1 Accuracy (%)76.68Unverified
10resnet8x4 (T: resnet32x4 S: resnet8x4)Top-1 Accuracy (%)76.31Unverified
#ModelMetricClaimedVerifiedStatus
1LSHFM (T: ResNet101 S: ResNet50)mAP93.17Unverified
2LSHFM (T: ResNet101 S: MobileNetV2)mAP90.14Unverified
#ModelMetricClaimedVerifiedStatus
1TIE-KD (T: Adabins S: MobileNetV2)RMSE2.43Unverified