SOTAVerified

Multimodal Autoregressive Pre-training of Large Vision Encoders

2024-11-21CVPR 2025Code Available5· sign in to hype

Enrico Fini, Mustafa Shukor, Xiujun Li, Philipp Dufter, Michal Klein, David Haldimann, Sai Aitharaju, Victor Guilherme Turrisi da Costa, Louis Béthune, Zhe Gan, Alexander T Toshev, Marcin Eichner, Moin Nabi, Yinfei Yang, Joshua M. Susskind, Alaaeldin El-Nouby

Code Available — Be the first to reproduce this paper.

Reproduce

Code

Abstract

We introduce a novel method for pre-training of large-scale vision encoders. Building on recent advancements in autoregressive pre-training of vision models, we extend this framework to a multimodal setting, i.e., images and text. In this paper, we present AIMV2, a family of generalist vision encoders characterized by a straightforward pre-training process, scalability, and remarkable performance across a range of downstream tasks. This is achieved by pairing the vision encoder with a multimodal decoder that autoregressively generates raw image patches and text tokens. Our encoders excel not only in multimodal evaluations but also in vision benchmarks such as localization, grounding, and classification. Notably, our AIMV2-3B encoder achieves 89.5% accuracy on ImageNet-1k with a frozen trunk. Furthermore, AIMV2 consistently outperforms state-of-the-art contrastive models (e.g., CLIP, SigLIP) in multimodal image understanding across diverse settings.

Tasks

Benchmark Results

DatasetModelMetricClaimedVerifiedStatus
ImageNetAIMv2-LTop 1 Accuracy86.6Unverified
ImageNetAIMv2-3B (448 res)Top 1 Accuracy89.5Unverified
ImageNetAIMv2-3BTop 1 Accuracy88.5Unverified
ImageNetAIMv2-1BTop 1 Accuracy88.1Unverified
ImageNetAIMv2-HTop 1 Accuracy87.5Unverified
iNaturalistAIMv2-3B (448 res)Top 1 Accuracy85.9Unverified
iNaturalistAIMv2-3BTop 1 Accuracy81.5Unverified
iNaturalistAIMv2-1BTop 1 Accuracy79.7Unverified
iNaturalistAIMv2-HTop 1 Accuracy77.9Unverified
iNaturalistAIMv2-LTop 1 Accuracy76Unverified

Reproductions