SOTAVerified

MMRL: Multi-Modal Representation Learning for Vision-Language Models

2025-03-11CVPR 2025Code Available2· sign in to hype

Yuncheng Guo, Xiaodong Gu

Code Available — Be the first to reproduce this paper.

Reproduce

Code

Abstract

Large-scale pre-trained Vision-Language Models (VLMs) have become essential for transfer learning across diverse tasks. However, adapting these models with limited few-shot data often leads to overfitting, diminishing their performance on new tasks. To tackle this issue, we propose a novel Multi-Modal Representation Learning (MMRL) framework that introduces a shared, learnable, and modality-agnostic representation space. MMRL projects the space tokens to text and image representation tokens, facilitating more effective multi-modal interactions. Unlike previous approaches that solely optimize class token features, MMRL integrates representation tokens at higher layers of the encoders--where dataset-specific features are more prominent--while preserving generalized knowledge in the lower layers. During training, both representation and class features are optimized, with trainable projection layer applied to the representation tokens, whereas the class token projection layer remains frozen to retain pre-trained knowledge. Furthermore, a regularization term is introduced to align the class features and text features with the zero-shot features from the frozen VLM, thereby safeguarding the model's generalization capacity. For inference, a decoupling strategy is employed, wherein both representation and class features are utilized for base classes, while only the class features, which retain more generalized knowledge, are used for new tasks. Extensive experiments across 15 datasets demonstrate that MMRL outperforms state-of-the-art methods, achieving a balanced trade-off between task-specific adaptation and generalization. Code is available at https://github.com/yunncheng/MMRL.

Tasks

Benchmark Results

DatasetModelMetricClaimedVerifiedStatus
Caltech-101MMRLHarmonic mean96.68Unverified
DTDMMRLHarmonic mean73.82Unverified
EuroSATMMRLHarmonic mean87.21Unverified
FGVC-AircraftMMRLHarmonic mean41.15Unverified
Food-101MMRLHarmonic mean91.03Unverified
ImageNetMMRLHarmonic mean74.45Unverified
ImageNet-AMMRLTop-1 accuracy %51.2Unverified
ImageNet-RMMRLTop-1 accuracy %77.53Unverified
ImageNet-SMMRLTop-1 accuracy %49.17Unverified
ImageNet V2MMRLTop-1 accuracy %64.47Unverified
Oxford 102 FlowerMMRLHarmonic mean86.78Unverified
Oxford-IIIT Pet DatasetMMRLHarmonic mean96.74Unverified
Stanford CarsMMRLHarmonic mean78.06Unverified
SUN397MMRLHarmonic mean81.2Unverified
UCF101MMRLHarmonic mean83.89Unverified

Reproductions