SOTAVerified

Knowledge Distillation

Knowledge distillation is the process of transferring knowledge from a large model to a smaller one. While large models (such as very deep neural networks or ensembles of many models) have higher knowledge capacity than small models, this capacity might not be fully utilized.

Papers

Showing 23512400 of 4240 papers

TitleStatusHype
FEED: Feature-level Ensemble for Knowledge Distillation0
Few-shot 3D LiDAR Semantic Segmentation for Autonomous Driving0
Few-shot Face Image Translation via GAN Prior Distillation0
Few-shot learning of neural networks from scratch by pseudo example optimization0
Few-Shot Object Detection by Knowledge Distillation Using Bag-of-Visual-Words Representations0
Optimizing Vision Transformers with Data-Free Knowledge Transfer0
Optimizing YOLOv5s Object Detection through Knowledge Distillation algorithm0
Oracle Teacher: Leveraging Target Information for Better Knowledge Distillation of CTC Models0
Orderly Dual-Teacher Knowledge Distillation for Lightweight Human Pose Estimation0
ORQA: A Benchmark and Foundation Model for Holistic Operating Room Modeling0
Overcoming Language Priors for Visual Question Answering Based on Knowledge Distillation0
P4: Towards private, personalized, and Peer-to-Peer learning0
Pacemaker: Intermediate Teacher Knowledge Distillation For On-The-Fly Convolutional Neural Network0
PAIR: Leveraging Passage-Centric Similarity Relation for Improving Dense Passage Retrieval0
Pan-infection Foundation Framework Enables Multiple Pathogen Prediction0
PANLP at MEDIQA 2019: Pre-trained Language Models, Transfer Learning and Knowledge Distillation0
Papago’s Submission for the WMT21 Quality Estimation Shared Task0
Paralinguistic Privacy Protection at the Edge0
Parameter-Efficient and Student-Friendly Knowledge Distillation0
Parameter-Efficient Conformers via Sharing Sparsely-Gated Experts for End-to-End Speech Recognition0
Parameter Efficient Diverse Paraphrase Generation Using Sequence-Level Knowledge Distillation0
Partial Knowledge Distillation for Alleviating the Inherent Inter-Class Discrepancy in Federated Learning0
Partial to Whole Knowledge Distillation: Progressive Distilling Decomposed Knowledge Boosts Student Better0
PC-LoRA: Low-Rank Adaptation for Progressive Model Compression with Knowledge Distillation0
PDALN: Progressive Domain Adaptation over a Pre-trained Model for Low-Resource Cross-Domain Named Entity Recognition0
Peak-Controlled Logits Poisoning Attack in Federated Distillation0
Pea-KD: Parameter-efficient and Accurate Knowledge Distillation on BERT0
Pea-KD: Parameter-efficient and accurate Knowledge Distillation0
Peak-First CTC: Reducing the Peak Latency of CTC Models by Applying Peak-First Regularization0
Peer Collaborative Learning for Polyphonic Sound Event Detection0
Learning to Maximize Speech Quality Directly Using MOS Prediction for Neural Text-to-Speech0
Performance-Aware Mutual Knowledge Distillation for Improving Neural Architecture Search0
Performance-Efficiency Trade-Offs in Adapting Language Models to Text Classification Tasks0
Performance-Guided LLM Knowledge Distillation for Efficient Text Classification at Scale0
Periocular Embedding Learning with Consistent Knowledge Distillation from Face0
Personalised Federated Learning: A Combinational Approach0
Personalized Decentralized Federated Learning with Knowledge Distillation0
PGX: A Multi-level GNN Explanation Framework Based on Separate Knowledge Distillation Processes0
PHI-S: Distribution Balancing for Label-Free Multi-Teacher Distillation0
PicoSAM2: Low-Latency Segmentation In-Sensor for Edge Vision Applications0
PILE: Pairwise Iterative Logits Ensemble for Multi-Teacher Labeled Distillation0
PIRB: A Comprehensive Benchmark of Polish Dense and Hybrid Text Retrieval Methods0
PISCO: Pretty Simple Compression for Retrieval-Augmented Generation0
Pixel Invisibility: Detecting Objects Invisible in Color Images0
P-KDGAN: Progressive Knowledge Distillation with GANs for One-class Novelty Detection0
PKD: General Distillation Framework for Object Detectors via Pearson Correlation Coefficient0
PLaD: Preference-based Large Language Model Distillation with Pseudo-Preference Pairs0
PlaStIL: Plastic and Stable Memory-Free Class-Incremental Learning0
Plug-and-Play Interpretable Responsible Text-to-Image Generation via Dual-Space Multi-facet Concept Control0
Point Adversarial Self Mining: A Simple Method for Facial Expression Recognition0
Show:102550
← PrevPage 48 of 85Next →

Benchmark Results

#ModelMetricClaimedVerifiedStatus
1ScaleKD (T:BEiT-L S:ViT-B/14)Top-1 accuracy %86.43Unverified
2ScaleKD (T:Swin-L S:ViT-B/16)Top-1 accuracy %85.53Unverified
3ScaleKD (T:Swin-L S:ViT-S/16)Top-1 accuracy %83.93Unverified
4ScaleKD (T:Swin-L S:Swin-T)Top-1 accuracy %83.8Unverified
5KD++(T: regnety-16GF S:ViT-B)Top-1 accuracy %83.6Unverified
6VkD (T:RegNety 160 S:DeiT-S)Top-1 accuracy %82.9Unverified
7SpectralKD (T:Swin-S S:Swin-T)Top-1 accuracy %82.7Unverified
8ScaleKD (T:Swin-L S:ResNet-50)Top-1 accuracy %82.55Unverified
9DiffKD (T:Swin-L S: Swin-T)Top-1 accuracy %82.5Unverified
10DIST (T: Swin-L S: Swin-T)Top-1 accuracy %82.3Unverified
#ModelMetricClaimedVerifiedStatus
1SRD (T:resnet-32x4, S:shufflenet-v2)Top-1 Accuracy (%)79.86Unverified
2shufflenet-v2(T:resnet-32x4, S:shufflenet-v2)Top-1 Accuracy (%)78.76Unverified
3MV-MR (T: CLIP/ViT-B-16 S: resnet50)Top-1 Accuracy (%)78.6Unverified
4resnet8x4 (T: resnet32x4 S: resnet8x4)Top-1 Accuracy (%)78.28Unverified
5resnet8x4 (T: resnet32x4 S: resnet8x4 [modified])Top-1 Accuracy (%)78.08Unverified
6ReviewKD++(T:resnet-32x4, S:shufflenet-v2)Top-1 Accuracy (%)77.93Unverified
7ReviewKD++(T:resnet-32x4, S:shufflenet-v1)Top-1 Accuracy (%)77.68Unverified
8resnet8x4 (T: resnet32x4 S: resnet8x4)Top-1 Accuracy (%)77.5Unverified
9resnet8x4 (T: resnet32x4 S: resnet8x4)Top-1 Accuracy (%)76.68Unverified
10resnet8x4 (T: resnet32x4 S: resnet8x4)Top-1 Accuracy (%)76.31Unverified
#ModelMetricClaimedVerifiedStatus
1LSHFM (T: ResNet101 S: ResNet50)mAP93.17Unverified
2LSHFM (T: ResNet101 S: MobileNetV2)mAP90.14Unverified
#ModelMetricClaimedVerifiedStatus
1TIE-KD (T: Adabins S: MobileNetV2)RMSE2.43Unverified