SOTAVerified

Meta-Learning

Meta-learning is a methodology considered with "learning to learn" machine learning algorithms.

( Image credit: Model-Agnostic Meta-Learning for Fast Adaptation of Deep Networks )

Papers

Showing 651675 of 3569 papers

TitleStatusHype
A Complete Survey on Contemporary Methods, Emerging Paradigms and Hybrid Approaches for Few-Shot Learning0
Automatic Combination of Sample Selection Strategies for Few-Shot Learning0
Symbol: Generating Flexible Black-Box Optimizers through Symbolic Equation LearningCode1
Rethinking the Starting Point: Collaborative Pre-Training for Federated Downstream Tasks0
Human-like Category Learning by Injecting Ecological Priors from Large Language Models into Neural Networks0
A Survey of Few-Shot Learning on Graphs: from Meta-Learning to Pre-Training and Prompt LearningCode1
CPT: Competence-progressive Training Strategy for Few-shot Node Classification0
Sample Weight Estimation Using Meta-Updates for Online Continual LearningCode0
Meta-Learning for Neural Network-based Temporal Point Processes0
An Information-Theoretic Analysis of In-Context Learning0
Proto-MPC: An Encoder-Prototype-Decoder Approach for Quadrotor Control in Challenging Winds0
Learning Universal PredictorsCode2
Meta-Learning Linear Quadratic Regulators: A Policy Gradient MAML Approach for Model-free LQR0
Incorporating Test-Time Optimization into Training with Dual Networks for Human Mesh RecoveryCode0
A Cost-Sensitive Meta-Learning Strategy for Fair Provider Exposure in RecommendationCode0
CDRNP: Cross-Domain Recommendation to Cold-Start Users via Neural Process0
MetaSeg: Content-Aware Meta-Net for Omni-Supervised Semantic Segmentation0
SAGE-HB: Swift Adaptation and Generalization in Massive MIMO Hybrid Beamforming0
DK-SLAM: Monocular Visual SLAM with Deep Keypoint Learning, Tracking and Loop-Closing0
Efficient Neural Representation of Volumetric Data using Coordinate-Based Networks0
Multi-view Distillation based on Multi-modal Fusion for Few-shot Action Recognition(CLIP-M^2DF)Code0
Only Send What You Need: Learning to Communicate Efficiently in Federated Multilingual Machine Translation0
Fine-Grained Prototypes Distillation for Few-Shot Object DetectionCode2
Window Stacking Meta-Models for Clinical EEG ClassificationCode0
Secrets of RLHF in Large Language Models Part II: Reward ModelingCode5
Show:102550
← PrevPage 27 of 143Next →

Benchmark Results

#ModelMetricClaimedVerifiedStatus
1MZ+ReconMeta-train success rate97.8Unverified
2MZMeta-train success rate97.6Unverified
3MAMLMeta-test success rate36Unverified
4RL^2Meta-test success rate10Unverified
5DnCMeta-test success rate5.4Unverified
6PEARLMeta-test success rate0Unverified
#ModelMetricClaimedVerifiedStatus
1SoftModuleAverage Success Rate60Unverified
2Multi-task multi-head SACAverage Success Rate35.85Unverified
3DisCorAverage Success Rate26Unverified
4NDPAverage Success Rate11Unverified
#ModelMetricClaimedVerifiedStatus
1MZ+ReconMeta-test success rate (zero-shot)18.5Unverified
2MZMeta-test success rate (zero-shot)17.7Unverified
#ModelMetricClaimedVerifiedStatus
1Metadrop% Test Accuracy95.75Unverified