SOTAVerified

MMRL++: Parameter-Efficient and Interaction-Aware Representation Learning for Vision-Language Models

2025-05-15Code Available2· sign in to hype

Yuncheng Guo, Xiaodong Gu

Code Available — Be the first to reproduce this paper.

Reproduce

Code

Abstract

Large-scale pre-trained Vision-Language Models (VLMs) have significantly advanced transfer learning across diverse tasks. However, adapting these models with limited few-shot data often leads to overfitting, undermining their ability to generalize to new tasks. To address this, we propose Multi-Modal Representation Learning (MMRL), which introduces a shared, learnable, modality-agnostic representation space. MMRL generates space tokens projected into both text and image encoders as representation tokens, enabling more effective cross-modal interactions. Unlike prior methods that mainly optimize class token features, MMRL inserts representation tokens into higher encoder layers--where task-specific features are more prominent--while preserving general knowledge in the lower layers. During training, both class and representation features are jointly optimized: a trainable projection layer is applied to representation tokens for task adaptation, while the projection layer for class token remains frozen to retain pre-trained knowledge. To further promote generalization, we introduce a regularization term aligning class and text features with the frozen VLM's zero-shot features. At inference, a decoupling strategy uses both class and representation features for base tasks, but only class features for novel tasks due to their stronger generalization. Building upon this, we propose MMRL++, a parameter-efficient and interaction-aware extension that significantly reduces trainable parameters and enhances intra-modal interactions--particularly across the layers of representation tokens--allowing gradient sharing and instance-specific information to propagate more effectively through the network. Extensive experiments on 15 datasets demonstrate that MMRL and MMRL++ consistently outperform state-of-the-art methods, achieving a strong balance between task-specific adaptation and generalization.

Tasks

Benchmark Results

DatasetModelMetricClaimedVerifiedStatus
Caltech-101MMRL++Harmonic mean96.75Unverified
DTDMMRL++Harmonic mean74.46Unverified
EuroSATMMRL++Harmonic mean91.94Unverified
FGVC-AircraftMMRL++Harmonic mean42.24Unverified
Food-101MMRL++Harmonic mean91.1Unverified
ImageNetMMRL++Harmonic mean74.44Unverified
Oxford 102 FlowerMMRL++Harmonic mean87.01Unverified
Oxford-IIIT Pet DatasetMMRL++Harmonic mean96.51Unverified
Stanford CarsMMRL++Harmonic mean78.18Unverified
SUN397MMRL++Harmonic mean81.28Unverified
UCF101MMRL++Harmonic mean83.81Unverified

Reproductions