论文标题
时间序列变压器生成对抗网络
Time-series Transformer Generative Adversarial Networks
论文作者
论文摘要
许多现实世界的任务受到数据限制的困扰:在某些情况下,数据很少,在其他情况下,数据受到隐私执行法规的保护(例如GDPR)。我们考虑在时间序列数据上特别提出的局限性,并提出一个模型,该模型可以生成可以代替实际数据的合成时间序列。生成合成时间序列数据的模型具有两个目标:1)捕获真实序列的逐步条件分布,以及2)忠实地对整个真实序列的关节分布进行建模。通过最大似然估计训练的自回归模型可以在以前的预测后被馈入并用于预测未来的模型。在这样的模型中,随着时间的流逝,错误可能会产生。此外,需要一个合理的初始值,使基于MLE的模型并非真正生成。许多下游任务学会学会建模时间序列的条件分布,因此,从生成模型中绘制的合成数据还必须满足1)除了执行2)。我们提出了TST-GAN,这是一个框架,该框架利用了变压器体系结构,以满足Desiderata并将其性能与五个数据集上的五个最先进的模型进行比较,并表明TST-GAN在所有数据集中都能达到更高的预测性能。
Many real-world tasks are plagued by limitations on data: in some instances very little data is available and in others, data is protected by privacy enforcing regulations (e.g. GDPR). We consider limitations posed specifically on time-series data and present a model that can generate synthetic time-series which can be used in place of real data. A model that generates synthetic time-series data has two objectives: 1) to capture the stepwise conditional distribution of real sequences, and 2) to faithfully model the joint distribution of entire real sequences. Autoregressive models trained via maximum likelihood estimation can be used in a system where previous predictions are fed back in and used to predict future ones; in such models, errors can accrue over time. Furthermore, a plausible initial value is required making MLE based models not really generative. Many downstream tasks learn to model conditional distributions of the time-series, hence, synthetic data drawn from a generative model must satisfy 1) in addition to performing 2). We present TsT-GAN, a framework that capitalises on the Transformer architecture to satisfy the desiderata and compare its performance against five state-of-the-art models on five datasets and show that TsT-GAN achieves higher predictive performance on all datasets.