SOTAVerified

Quantum Algorithm for Sparse Online Learning with Truncated Gradient Descent

2024-11-06Unverified0· sign in to hype

Debbie Lim, Yixian Qiu, Patrick Rebentrost, Qisheng Wang

Unverified — Be the first to reproduce this paper.

Reproduce

Abstract

Logistic regression, the Support Vector Machine (SVM), and least squares are well-studied methods in the statistical and computer science community, with various practical applications. High-dimensional data arriving on a real-time basis makes the design of online learning algorithms that produce sparse solutions essential. The seminal work of cite.langford2009sparseLangford, Li, and Zhang (2009) developed a method to obtain sparsity via truncated gradient descent, showing a near-optimal online regret bound. Based on this method, we develop a quantum sparse online learning algorithm for logistic regression, the SVM, and least squares. Given efficient quantum access to the inputs, we show that a quadratic speedup in the time complexity with respect to the dimension of the problem is achievable, while maintaining a regret of O(1/T), where T is the number of iterations.

Tasks

Reproductions