SOTAVerified

Sparse Dimensionality Reduction Revisited

2023-02-13Unverified0· sign in to hype

Mikael Møller Høgsgaard, Lion Kamma, Kasper Green Larsen, Jelani Nelson, Chris Schwiegelshohn

Unverified — Be the first to reproduce this paper.

Reproduce

Abstract

The sparse Johnson-Lindenstrauss transform is one of the central techniques in dimensionality reduction. It supports embedding a set of n points in R^d into m=O(^-2 n) dimensions while preserving all pairwise distances to within 1 . Each input point x is embedded to Ax, where A is an m d matrix having s non-zeros per column, allowing for an embedding time of O(s \|x\|_0). Since the sparsity of A governs the embedding time, much work has gone into improving the sparsity s. The current state-of-the-art by Kane and Nelson (JACM'14) shows that s = O( ^-1 n) suffices. This is almost matched by a lower bound of s = ( ^-1 n/(1/)) by Nelson and Nguyen (STOC'13). Previous work thus suggests that we have near-optimal embeddings. In this work, we revisit sparse embeddings and identify a loophole in the lower bound. Concretely, it requires d n, which in many applications is unrealistic. We exploit this loophole to give a sparser embedding when d = o(n), achieving s = O(^-1( n/(1/)+^2/3n ^1/3 d)). We also complement our analysis by strengthening the lower bound of Nelson and Nguyen to hold also when d n, thereby matching the first term in our new sparsity upper bound. Finally, we also improve the sparsity of the best oblivious subspace embeddings for optimal embedding dimensionality.

Tasks

Reproductions