SOTAVerified

Subgradient Descent Learns Orthogonal Dictionaries

2018-10-25ICLR 2019Code Available0· sign in to hype

Yu Bai, Qijia Jiang, Ju Sun

Code Available — Be the first to reproduce this paper.

Reproduce

Code

Abstract

This paper concerns dictionary learning, i.e., sparse coding, a fundamental representation learning problem. We show that a subgradient descent algorithm, with random initialization, can provably recover orthogonal dictionaries on a natural nonsmooth, nonconvex _1 minimization formulation of the problem, under mild statistical assumptions on the data. This is in contrast to previous provable methods that require either expensive computation or delicate initialization schemes. Our analysis develops several tools for characterizing landscapes of nonsmooth functions, which might be of independent interest for provable training of deep networks with nonsmooth activations (e.g., ReLU), among numerous other applications. Preliminary experiments corroborate our analysis and show that our algorithm works well empirically in recovering orthogonal dictionaries.

Tasks

Reproductions