Pretraining Finnish ModernBERTs
2025-11-12Code Available0· sign in to hype
Akseli Reunamo, Laura-Maria Peltonen, Hans Moen, Sampo Pyysalo
Code Available — Be the first to reproduce this paper.
ReproduceCode
- github.com/rakseli/finnish-modernbertsOfficial★ 2
Abstract
This paper reports on pretraining ModernBERT encoder models in six different sizes, ranging from 51M to 475M parameters, with a focus on limited multilingualism, emphasizing languages relevant to Finland. Our models are competitive with, or superior to, existing multilingual models. They outperform monolingual models on tasks that require a context longer than 512 tokens. We present empirical results on using different data in the final stage of training. The code and models are publicly released.