SOTAVerified

UniPrompt-CL: Sustainable Continual Learning in Medical AI with Unified Prompt Pools

2026-03-13Unverified0· sign in to hype

Gyutae Oh, Jitae Shin

Unverified — Be the first to reproduce this paper.

Reproduce

Abstract

Modern AI models are typically trained on static datasets, limiting their ability to continuously adapt to rapidly evolving real-world environments. While continual learning (CL) addresses this limitation, most CL methods are designed for natural images and often underperform or fail to transfer to medical data due to domain bias, institutional constraints, and subtle inter-stage boundaries. We propose UniPrompt-CL, a medical-oriented prompt-based continual learning method that improves prompt pool design via a minimally expanding unified prompt pool and a new regularization term, achieving a better stability-plasticity trade-off with lower computational cost. Across two domain-incremental learning settings, UniPrompt-CL effectively reduces inference cost while improving AvgACC by 1-3 percentage points. In addition to strong performance, extensive experiments clearly validate the motivation and effectiveness of the proposed improvements.

Reproductions