SOTAVerified

Continual Pretraining

Papers

Showing 2650 of 70 papers

TitleStatusHype
Robust Data Watermarking in Language Models by Injecting Fictitious KnowledgeCode0
RomanSetu: Efficiently unlocking multilingual capabilities of Large Language Models via RomanizationCode0
Towards Democratizing Multilingual Large Language Models For Medicine Through A Two-Stage Instruction Fine-tuning ApproachCode0
Fortunately, Discourse Markers Can Enhance Language Models for Sentiment AnalysisCode0
Simulating Training Data Leakage in Multiple-Choice Benchmarks for LLM EvaluationCode0
Hierarchical Label-wise Attention Transformer Model for Explainable ICD CodingCode0
Unsupervised Domain Adaptation for Sparse Retrieval by Filling Vocabulary and Word Frequency GapsCode0
PECoP: Parameter Efficient Continual Pretraining for Action Quality AssessmentCode0
AF Adapter: Continual Pretraining for Building Chinese Biomedical Language ModelCode0
LangSAMP: Language-Script Aware Multilingual PretrainingCode0
Alchemy: Amplifying Theorem-Proving Capability through Symbolic MutationCode0
PARAMANU-AYN: Pretrain from scratch or Continual Pretraining of LLMs for Legal Domain Adaptation?0
Pretraining and Updates of Domain-Specific LLM: A Case Study in the Japanese Business Domain0
RedWhale: An Adapted Korean LLM Through Efficient Continual Pretraining0
Revisiting Pretraining with Adapters0
AfroXLMR-Social: Adapting Pre-trained Language Models for African Languages Social Media Text0
The Construction of Instruction-tuned LLMs for Finance without Instruction Data Using Continual Pretraining and Model Merging0
AdaPrompt: Adaptive Model Training for Prompt-based NLP0
BAMBINO-LM: (Bilingual-)Human-Inspired Continual Pretraining of BabyLM0
Bilingual Adaptation of Monolingual Foundation Models0
Breaking the Stage Barrier: A Novel Single-Stage Approach to Long Context Extension for Large Language Models0
Biomed-Enriched: A Biomedical Dataset Enriched with LLMs for Pretraining and Extracting Rare and Hidden Content0
ChuXin: 1.6B Technical Report0
Continual Learning for Large Language Models: A Survey0
70B-parameter large language models in Japanese medical question-answering0
Show:102550
← PrevPage 2 of 3Next →

Benchmark Results

#ModelMetricClaimedVerifiedStatus
1DASF1 (macro)0.69Unverified
#ModelMetricClaimedVerifiedStatus
1CPTF1 - macro63.77Unverified
#ModelMetricClaimedVerifiedStatus
1DASF1 (macro)0.71Unverified