SOTAVerified

SutraNets: Sub-series Autoregressive Networks for Long-Sequence, Probabilistic Forecasting

2023-12-22NeurIPS 2023Unverified0· sign in to hype

Shane Bergsma, Timothy Zeyl, Lei Guo

Unverified — Be the first to reproduce this paper.

Reproduce

Abstract

We propose SutraNets, a novel method for neural probabilistic forecasting of long-sequence time series. SutraNets use an autoregressive generative model to factorize the likelihood of long sequences into products of conditional probabilities. When generating long sequences, most autoregressive approaches suffer from harmful error accumulation, as well as challenges in modeling long-distance dependencies. SutraNets treat long, univariate prediction as multivariate prediction over lower-frequency sub-series. Autoregression proceeds across time and across sub-series in order to ensure coherent multivariate (and, hence, high-frequency univariate) outputs. Since sub-series can be generated using fewer steps, SutraNets effectively reduce error accumulation and signal path distances. We find SutraNets to significantly improve forecasting accuracy over competitive alternatives on six real-world datasets, including when we vary the number of sub-series and scale up the depth and width of the underlying sequence models.

Tasks

Reproductions