SOTAVerified

Read-only Prompt Optimization for Vision-Language Few-shot Learning

2023-08-29ICCV 2023Code Available1· sign in to hype

Dongjun Lee, Seokwon Song, Jihee Suh, Joonmyung Choi, Sanghyeok Lee, Hyunwoo J. Kim

Code Available — Be the first to reproduce this paper.

Reproduce

Code

Abstract

In recent years, prompt tuning has proven effective in adapting pre-trained vision-language models to downstream tasks. These methods aim to adapt the pre-trained models by introducing learnable prompts while keeping pre-trained weights frozen. However, learnable prompts can affect the internal representation within the self-attention module, which may negatively impact performance variance and generalization, especially in data-deficient settings. To address these issues, we propose a novel approach, Read-only Prompt Optimization (RPO). RPO leverages masked attention to prevent the internal representation shift in the pre-trained model. Further, to facilitate the optimization of RPO, the read-only prompts are initialized based on special tokens of the pre-trained model. Our extensive experiments demonstrate that RPO outperforms CLIP and CoCoOp in base-to-new generalization and domain generalization while displaying better robustness. Also, the proposed method achieves better generalization on extremely data-deficient settings, while improving parameter efficiency and computational overhead. Code is available at https://github.com/mlvlab/RPO.

Tasks

Benchmark Results

DatasetModelMetricClaimedVerifiedStatus
Caltech-101RPOHarmonic mean96.03Unverified
DTDRPOHarmonic mean68.61Unverified
EuroSATRPOHarmonic mean76.79Unverified
FGVC-AircraftRPOHarmonic mean35.7Unverified
Food-101RPOHarmonic mean90.58Unverified
ImageNetRPOHarmonic mean74Unverified
Oxford 102 FlowerRPOHarmonic mean84.5Unverified
Oxford-IIIT Pet DatasetRPOHarmonic mean96.05Unverified
Stanford CarsRPOHarmonic mean74.69Unverified
SUN397RPOHarmonic mean79.18Unverified
UCF101RPOHarmonic mean79.34Unverified

Reproductions