SOTAVerified

LoRA Training in the NTK Regime has No Spurious Local Minima

2024-02-19Code Available0· sign in to hype

Uijeong Jang, Jason D. Lee, Ernest K. Ryu

Code Available — Be the first to reproduce this paper.

Reproduce

Code

Abstract

Low-rank adaptation (LoRA) has become the standard approach for parameter-efficient fine-tuning of large language models (LLM), but our theoretical understanding of LoRA has been limited. In this work, we theoretically analyze LoRA fine-tuning in the neural tangent kernel (NTK) regime with N data points, showing: (i) full fine-tuning (without LoRA) admits a low-rank solution of rank r N; (ii) using LoRA with rank r N eliminates spurious local minima, allowing gradient descent to find the low-rank solutions; (iii) the low-rank solution found using LoRA generalizes well.

Tasks

Reproductions