SOTAVerified

Non-Uniform Parameter-Wise Model Merging

2024-12-20Unverified0· sign in to hype

Albert Manuel Orozco Camacho, Stefan Horoi, Guy Wolf, Eugene Belilovsky

Unverified — Be the first to reproduce this paper.

Reproduce

Abstract

Combining multiple machine learning models has long been a technique for enhancing performance, particularly in distributed settings. Traditional approaches, such as model ensembles, work well, but are expensive in terms of memory and compute. Recently, methods based on averaging model parameters have achieved good results in some settings and have gained popularity. However, merging models initialized differently that do not share a part of their training trajectories can yield worse results than simply using the base models, even after aligning their neurons. In this paper, we introduce a novel approach, Non-uniform Parameter-wise Model Merging, or NP Merge, which merges models by learning the contribution of each parameter to the final model using gradient-based optimization. We empirically demonstrate the effectiveness of our method for merging models of various architectures in multiple settings, outperforming past methods. We also extend NP Merge to handle the merging of multiple models, showcasing its scalability and robustness.

Tasks

Reproductions