SOTAVerified

Continual Pretraining

Papers

Showing 3140 of 70 papers

TitleStatusHype
Towards Lifelong Learning of Large Language Models: A SurveyCode2
LongSkywork: A Training Recipe for Efficiently Extending Context Length in Large Language Models0
Multi-Label Guided Soft Contrastive Learning for Efficient Earth Observation PretrainingCode1
MoRA: High-Rank Updating for Parameter-Efficient Fine-TuningCode3
Cross-sensor self-supervised training and alignment for remote sensing0
ChuXin: 1.6B Technical Report0
Retrieval Head Mechanistically Explains Long-Context FactualityCode3
Pretraining and Updates of Domain-Specific LLM: A Case Study in the Japanese Business Domain0
CEM: A Data-Efficient Method for Large Language Models to Continue Evolving From Mistakes0
Rho-1: Not All Tokens Are What You NeedCode3
Show:102550
← PrevPage 4 of 7Next →

Benchmark Results

#ModelMetricClaimedVerifiedStatus
1DASF1 (macro)0.69Unverified
#ModelMetricClaimedVerifiedStatus
1CPTF1 - macro63.77Unverified
#ModelMetricClaimedVerifiedStatus
1DASF1 (macro)0.71Unverified