SOTAVerified

Scaffold with Stochastic Gradients: New Analysis with Linear Speed-Up

2025-03-10Code Available0· sign in to hype

Paul Mangold, Alain Durmus, Aymeric Dieuleveut, Eric Moulines

Code Available — Be the first to reproduce this paper.

Reproduce

Code

Abstract

This paper proposes a novel analysis for the Scaffold algorithm, a popular method for dealing with data heterogeneity in federated learning. While its convergence in deterministic settings--where local control variates mitigate client drift--is well established, the impact of stochastic gradient updates on its performance is less understood. To address this problem, we first show that its global parameters and control variates define a Markov chain that converges to a stationary distribution in the Wasserstein distance. Leveraging this result, we prove that Scaffold achieves linear speed-up in the number of clients up to higher-order terms in the step size. Nevertheless, our analysis reveals that Scaffold retains a higher-order bias, similar to FedAvg, that does not decrease as the number of clients increases. This highlights opportunities for developing improved stochastic federated learning algorithms

Tasks

Reproductions