SPIDER: Near-Optimal Non-Convex Optimization via Stochastic Path Integrated Differential Estimator
Cong Fang, Chris Junchi Li, Zhouchen Lin, Tong Zhang
Unverified — Be the first to reproduce this paper.
ReproduceAbstract
In this paper, we propose a new technique named Stochastic Path-Integrated Differential EstimatoR (SPIDER), which can be used to track many deterministic quantities of interest with significantly reduced computational cost. We apply SPIDER to two tasks, namely the stochastic first-order and zeroth-order methods. For stochastic first-order method, combining SPIDER with normalized gradient descent, we propose two new algorithms, namely SPIDER-SFO and SPIDER-SFO+, that solve non-convex stochastic optimization problems using stochastic gradients only. We provide sharp error-bound results on their convergence rates. In special, we prove that the SPIDER-SFO and SPIDER-SFO+ algorithms achieve a record-breaking gradient computation cost of O( ( n^1/2 ^-2, ^-3 ) ) for finding an -approximate first-order and O( ( n^1/2 ^-2+^-2.5, ^-3 ) ) for finding an (, O(^0.5))-approximate second-order stationary point, respectively. In addition, we prove that SPIDER-SFO nearly matches the algorithmic lower bound for finding approximate first-order stationary points under the gradient Lipschitz assumption in the finite-sum setting. For stochastic zeroth-order method, we prove a cost of O( d ( n^1/2 ^-2, ^-3) ) which outperforms all existing results.