【Paper】Transformers in Times Series: A Suervey

Abstract

In this paper, we sysytematically review Transformer schemes for time series modeling by hignlighting their strengths as well as limitations.
in particular, we examine the development of time series Transformers in two perspectives.

  • network structure. we sumarize the adaptation and modifacations that have been made to Transformers in order to accommodate the challenges. (我们总结了为了适应时间序列任务的各种挑战,而对 Transformers 所作的修改和调整)
  • applications: we categorize time series Transformer based on common tasks includeing foecasting, anomaly detection, and classfication.

Github: time-series-transformers-review

Introduction

Seasonality or periodicity is an effectively model long-range and short-range temporal dependency and capture seasonality simultaneously remains a chanllenge.

Preliminaries of the Transformer (Transforms 预备知识)

Vanilla Transformer

Vaila Transformer 解析

Input Encoding and Positional Encoding

Unlike LSTM or RNN, the vanilla Transformer has no recurrence. Instead, it utilizes the positional encoding added in the input embeddings, to model the sequence information. We summarize some positional encodings below:

Absuolute Positional Encoding

In vanilla Transformer, for each position index t t t , encoding vector is given by
在这里插入图片描述
where ω i \omega _i ωi is the had-crafted frequency for each dimension. Another way is to learn a set of positional embeddings for eachposition which is more flexibel.

Relative Positions Encoding
Multi-head Attention

Taxonomy of Transformers in Time Series

Figure 1

Network Modifications for Time Series

Positional Encoding

As the ordering of time series matters,(由于时间序列的顺序很重要), It is of great importance to encode the positions of input time series into Transformers. A common design is to first encode positional information as vector and then inject them into the model as an additional input together with the input time series. How to obtain these vectors when modeling time series with Transformers can be divided into thress main categories.

  • Vanilla positional Encoding: Unable to fully exploit the important features of time series data
  • Learnable Positional Encoding: As the vanilla positional encoding is hand-crafted and less expressive and adaptive ,several studies found that learning appropriate positional embeddings from time series data can be much more effective. Compared to fixed vanilla positional encoding, learned embeddings are more flexible and can adapt to specific tasks.
  • TimeStamp Encoding
  • 0
    点赞
  • 0
    收藏
    觉得还不错? 一键收藏
  • 0
    评论
评论
添加红包

请填写红包祝福语或标题

红包个数最小为10个

红包金额最低5元

当前余额3.43前往充值 >
需支付:10.00
成就一亿技术人!
领取后你会自动成为博主和红包主的粉丝 规则
hope_wisdom
发出的红包
实付
使用余额支付
点击重新获取
扫码支付
钱包余额 0

抵扣说明:

1.余额是钱包充值的虚拟货币,按照1:1的比例进行支付金额的抵扣。
2.余额无法直接购买下载,可以购买VIP、付费专栏及课程。

余额充值