SOTAVerified

R-LoRA: Random Initialization of Multi-Head LoRA for Multi-Task Learning

2025-02-21Code Available1· sign in to hype

Jinda Liu, Yi Chang, Yuan Wu

Code Available — Be the first to reproduce this paper.

Reproduce

Code

Abstract

Fine-tuning large language models (LLMs) is prohibitively expensive in terms of computational and memory costs. Low-rank Adaptation (LoRA), as one of the most popular parameter-efficient fine-tuning (PEFT) methods, offers a cost-effective alternative by approximating the model changes W R^m n through the product of down-projection matrix A R^m r and head matrix B R^r n, where r (m, n). In real-world scenarios, LLMs are fine-tuned on data from multiple domains to perform tasks across various fields, embodying multi-task learning (MTL). LoRA often underperforms in such complex scenarios. To enhance LoRA's capability in multi-task learning, we propose R-LoRA, which incorporates Multi-Head Randomization. Multi-Head Randomization diversifies the head matrices through Multi-Head Random Initialization and Multi-Head Dropout, enabling more efficient learning of task-specific features while maintaining shared knowledge representation. Extensive experiments demonstrate that R-LoRA is better at capturing task-specific knowledge, thereby improving performance in multi-task scenarios. The code is available at https://github.com/jinda-liu/R-LoRA.

Tasks

Reproductions