SOTAVerified

Informer: Beyond Efficient Transformer for Long Sequence Time-Series Forecasting

2020-12-14Code Available1· sign in to hype

Haoyi Zhou, Shanghang Zhang, Jieqi Peng, Shuai Zhang, JianXin Li, Hui Xiong, Wancai Zhang

Code Available — Be the first to reproduce this paper.

Reproduce

Code

Abstract

Many real-world applications require the prediction of long sequence time-series, such as electricity consumption planning. Long sequence time-series forecasting (LSTF) demands a high prediction capacity of the model, which is the ability to capture precise long-range dependency coupling between output and input efficiently. Recent studies have shown the potential of Transformer to increase the prediction capacity. However, there are several severe issues with Transformer that prevent it from being directly applicable to LSTF, including quadratic time complexity, high memory usage, and inherent limitation of the encoder-decoder architecture. To address these issues, we design an efficient transformer-based model for LSTF, named Informer, with three distinctive characteristics: (i) a ProbSparse self-attention mechanism, which achieves O(L L) in time complexity and memory usage, and has comparable performance on sequences' dependency alignment. (ii) the self-attention distilling highlights dominating attention by halving cascading layer input, and efficiently handles extreme long input sequences. (iii) the generative style decoder, while conceptually simple, predicts the long time-series sequences at one forward operation rather than a step-by-step way, which drastically improves the inference speed of long-sequence predictions. Extensive experiments on four large-scale datasets demonstrate that Informer significantly outperforms existing methods and provides a new solution to the LSTF problem.

Tasks

Benchmark Results

DatasetModelMetricClaimedVerifiedStatus
ETTh1 (168) MultivariateInformerMSE0.88Unverified
ETTh1 (168) UnivariateInformerMSE0.18Unverified
ETTh1 (24) MultivariateInformerMSE0.51Unverified
ETTh1 (24) UnivariateInformerMSE0.05Unverified
ETTh1 (336) MultivariateInformerMSE0.88Unverified
ETTh1 (336) UnivariateInformerMSE0.19Unverified
ETTh1 (48) MultivariateInformerMSE0.55Unverified
ETTh1 (48) UnivariateInformerMSE0.13Unverified
ETTh1 (720) MultivariateInformerMSE0.94Unverified
ETTh1 (720) UnivariateInformerMSE0.2Unverified
ETTh2 (168) MultivariateInformerMSE1.51Unverified
ETTh2 (168) UnivariateInformerMSE0.15Unverified
ETTh2 (24) MultivariateInformerMSE0.45Unverified
ETTh2 (24) UnivariateInformerMSE0.08Unverified
ETTh2 (336) MultivariateInformerMSE1.67Unverified
ETTh2 (336) UnivariateInformerMSE0.17Unverified
ETTh2 (48) MultivariateInformerMSE0.93Unverified
ETTh2 (48) UnivariateInformerMSE0.11Unverified
ETTh2 (720) MultivariateInformerMSE2.34Unverified
ETTh2 (720) UnivariateInformerMSE0.18Unverified

Reproductions