SOTAVerified

Continual Pretraining

Papers

Showing 110 of 70 papers

TitleStatusHype
Yi: Open Foundation Models by 01.AICode9
Scaling Granite Code Models to 128K ContextCode4
MoRA: High-Rank Updating for Parameter-Efficient Fine-TuningCode3
Retrieval Head Mechanistically Explains Long-Context FactualityCode3
Rho-1: Not All Tokens Are What You NeedCode3
Data Engineering for Scaling Language Models to 128K ContextCode3
A Practitioner's Guide to Continual Multimodal PretrainingCode2
Towards Lifelong Learning of Large Language Models: A SurveyCode2
Autonomous Data Selection with Zero-shot Generative Classifiers for Mathematical TextsCode2
Effective Long-Context Scaling of Foundation ModelsCode2
Show:102550
← PrevPage 1 of 7Next →

Benchmark Results

#ModelMetricClaimedVerifiedStatus
1DASF1 (macro)0.69Unverified
#ModelMetricClaimedVerifiedStatus
1CPTF1 - macro63.77Unverified
#ModelMetricClaimedVerifiedStatus
1DASF1 (macro)0.71Unverified