SOTAVerified

Pretraining Finnish ModernBERTs

2025-11-12Code Available0· sign in to hype

Akseli Reunamo, Laura-Maria Peltonen, Hans Moen, Sampo Pyysalo

Code Available — Be the first to reproduce this paper.

Reproduce

Code

Abstract

This paper reports on pretraining ModernBERT encoder models in six different sizes, ranging from 51M to 475M parameters, with a focus on limited multilingualism, emphasizing languages relevant to Finland. Our models are competitive with, or superior to, existing multilingual models. They outperform monolingual models on tasks that require a context longer than 512 tokens. We present empirical results on using different data in the final stage of training. The code and models are publicly released.

Reproductions