| Simulating Training Data Leakage in Multiple-Choice Benchmarks for LLM Evaluation | May 30, 2025 | Continual PretrainingFairness | CodeCode Available | 0 |
| A Japanese Language Model and Three New Evaluation Benchmarks for Pharmaceutical NLP | May 22, 2025 | Continual PretrainingDiagnostic | CodeCode Available | 0 |
| Enhance Mobile Agents Thinking Process Via Iterative Preference Learning | May 18, 2025 | Continual Pretraining | —Unverified | 0 |
| Mining Hidden Thoughts from Texts: Evaluating Continual Pretraining with Synthetic Data for LLM Reasoning | May 15, 2025 | Continual PretrainingMMLU | —Unverified | 0 |
| Efficient Domain-adaptive Continual Pretraining for the Process Industry in the German Language | Apr 28, 2025 | Continual PretrainingGPU | —Unverified | 0 |
| Enhancing Domain-Specific Encoder Models with LLM-Generated Data: How to Leverage Ontologies, and How to Do Without Them | Mar 27, 2025 | Continual PretrainingLanguage Modeling | —Unverified | 0 |
| Overcoming Vocabulary Mismatch: Vocabulary-agnostic Teacher Guided Language Modeling | Mar 24, 2025 | Continual PretrainingLanguage Modeling | —Unverified | 0 |
| AfroXLMR-Social: Adapting Pre-trained Language Models for African Languages Social Media Text | Mar 24, 2025 | Continual PretrainingEmotion Classification | —Unverified | 0 |
| Robust Data Watermarking in Language Models by Injecting Fictitious Knowledge | Mar 6, 2025 | Continual PretrainingMemorization | CodeCode Available | 0 |
| Multilingual Machine Translation with Open Large Language Models at Practical Scale: An Empirical Study | Feb 4, 2025 | Continual PretrainingMachine Translation | —Unverified | 0 |
| Breaking the Stage Barrier: A Novel Single-Stage Approach to Long Context Extension for Large Language Models | Dec 10, 2024 | Continual PretrainingLanguage Modeling | —Unverified | 0 |
| Alchemy: Amplifying Theorem-Proving Capability through Symbolic Mutation | Oct 21, 2024 | Automated Theorem ProvingContinual Pretraining | CodeCode Available | 0 |
| The Construction of Instruction-tuned LLMs for Finance without Instruction Data Using Continual Pretraining and Model Merging | Sep 30, 2024 | Continual Pretraining | —Unverified | 0 |
| DoPAMine: Domain-specific Pre-training Adaptation from seed-guided data Mining | Sep 30, 2024 | Continual PretrainingDomain Adaptation | —Unverified | 0 |
| AstroMLab 2: AstroLLaMA-2-70B Model and Benchmarking Specialised LLMs for Astronomy | Sep 29, 2024 | AstronomyBenchmarking | —Unverified | 0 |
| LangSAMP: Language-Script Aware Multilingual Pretraining | Sep 26, 2024 | Continual PretrainingLanguage Modeling | CodeCode Available | 0 |
| Towards Democratizing Multilingual Large Language Models For Medicine Through A Two-Stage Instruction Fine-tuning Approach | Sep 9, 2024 | Computational EfficiencyContinual Pretraining | CodeCode Available | 0 |
| RedWhale: An Adapted Korean LLM Through Efficient Continual Pretraining | Aug 21, 2024 | Continual PretrainingCross-Lingual Transfer | —Unverified | 0 |
| Bilingual Adaptation of Monolingual Foundation Models | Jul 13, 2024 | Continual PretrainingCross-Lingual Transfer | —Unverified | 0 |
| 70B-parameter large language models in Japanese medical question-answering | Jun 21, 2024 | Continual PretrainingDomain Adaptation | —Unverified | 0 |
| Open Generative Large Language Models for Galician | Jun 19, 2024 | Continual PretrainingDiversity | —Unverified | 0 |
| Understanding the RoPE Extensions of Long-Context LLMs: An Attention Perspective | Jun 19, 2024 | BenchmarkingContinual Pretraining | —Unverified | 0 |
| BAMBINO-LM: (Bilingual-)Human-Inspired Continual Pretraining of BabyLM | Jun 17, 2024 | Continual Pretrainingzero-shot-classification | —Unverified | 0 |
| LongSkywork: A Training Recipe for Efficiently Extending Context Length in Large Language Models | Jun 2, 2024 | Continual PretrainingInformation Retrieval | —Unverified | 0 |
| Cross-sensor self-supervised training and alignment for remote sensing | May 16, 2024 | Continual PretrainingEarth Observation | —Unverified | 0 |