SOTAVerified

Blending Low and High-Level Semantics of Time Series for Better Masked Time Series Generation

2024-08-29Unverified0· sign in to hype

Johan Vik Mathisen, Erlend Lokna, Daesoo Lee, Erlend Aune

Unverified — Be the first to reproduce this paper.

Reproduce

Abstract

State-of-the-art approaches in time series generation (TSG), such as TimeVQVAE, utilize vector quantization-based tokenization to effectively model complex distributions of time series. These approaches first learn to transform time series into a sequence of discrete latent vectors, and then a prior model is learned to model the sequence. The discrete latent vectors, however, only capture low-level semantics (e.g., shapes). We hypothesize that higher-fidelity time series can be generated by training a prior model on more informative discrete latent vectors that contain both low and high-level semantics (e.g., characteristic dynamics). In this paper, we introduce a novel framework, termed NC-VQVAE, to integrate self-supervised learning into those TSG methods to derive a discrete latent space where low and high-level semantics are captured. Our experimental results demonstrate that NC-VQVAE results in a considerable improvement in the quality of synthetic samples.

Tasks

Reproductions