TCCT: Tightly-coupled convolutional transformer on time series forecasting

被引:65
|
作者
Shen, Li [1 ]
Wang, Yangzhu [1 ]
机构
[1] Beihang Univ, Dayuncun Residential Quarter, RM 807,8th Dormitory,29 Zhichun Rd, Beijing 100191, Peoples R China
基金
中国国家自然科学基金;
关键词
Time series forecasting; Transformer; CNN;
D O I
10.1016/j.neucom.2022.01.039
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Time series forecasting is essential for a wide range of real-world applications. Recent studies have shown the superiority of Transformer in dealing with such problems, especially long sequence time series input (LSTI) and long sequence time series forecasting (LSTF) problems. To improve the efficiency and enhance the locality of Transformer, these studies combine Transformer with CNN in varying degrees. However, their combinations are loosely-coupled and do not make full use of CNN. To address this issue, we propose the concept of tightly-coupled convolutional Transformer (TCCT) and three TCCT architectures which apply transformed CNN architectures into Transformer: (1) CSPAttention: through fusing CSPNet with self-attention mechanism, the computation cost of self-attention mechanism is reduced by 30% and the memory usage is reduced by 50% while achieving equivalent or beyond prediction accuracy. (2) Dilated causal convolution: this method is to modify the distilling operation proposed by Informer through replacing canonical convolutional layers with dilated causal convolutional layers to gain exponentially receptive field growth. (3) Passthrough mechanism: the application of passthrough mechanism to stack of self-attention blocks helps Transformer-like models get more fine-grained information with negligible extra computation costs. Our experiments on real-world datasets show that our TCCT architectures could greatly improve the performance of existing state-of-the-art Transformer models on time series forecasting with much lower computation and memory costs, including canonical Transformer, LogTrans and Informer. (c) 2022 The Author(s). Published by Elsevier B.V. This is an open access article under the CC BY-NC-ND license (http://creativecommons.org/licenses/by-nc-nd/4.0/).
引用
收藏
页码:131 / 145
页数:15
相关论文
共 50 条
  • [1] ACT: Adversarial Convolutional Transformer for Time Series Forecasting
    Li, Yuan
    Wang, Huanjie
    Li, Jingwei
    Liu, Chengbao
    Tan, Jie
    2022 INTERNATIONAL JOINT CONFERENCE ON NEURAL NETWORKS (IJCNN), 2022,
  • [2] Exploiting Tightly-Coupled Cores
    Bates, Daniel
    Bradbury, Alex
    Koltes, Andreas
    Mullins, Robert
    JOURNAL OF SIGNAL PROCESSING SYSTEMS FOR SIGNAL IMAGE AND VIDEO TECHNOLOGY, 2015, 80 (01): : 103 - 120
  • [3] Exploiting Tightly-Coupled Cores
    Bates, Daniel
    Bradbury, Alex
    Koltes, Andreas
    Mullins, Robert
    2013 INTERNATIONAL CONFERENCE ON EMBEDDED COMPUTER SYSTEMS: ARCHITECTURES, MODELING AND SIMULATION (IC-SAMOS), 2013, : 296 - 305
  • [4] Exploiting Tightly-Coupled Cores
    Daniel Bates
    Alex Bradbury
    Andreas Koltes
    Robert Mullins
    Journal of Signal Processing Systems, 2015, 80 : 103 - 120
  • [5] DESIGN OF TIGHTLY-COUPLED MULTIPROCESSING PROGRAMMING
    ARNOLD, JS
    CASEY, DP
    MCKINSTRY, RH
    IBM SYSTEMS JOURNAL, 1974, 13 (01) : 60 - 87
  • [6] A tightly-coupled architecture for data mining
    Meo, R
    Psila, G
    Ceri, S
    14TH INTERNATIONAL CONFERENCE ON DATA ENGINEERING, PROCEEDINGS, 1998, : 316 - 323
  • [7] Tightly-coupled GPS/UWB Integration
    MacGougan, Glenn
    O'Keefe, Kyle
    Klukas, Richard
    JOURNAL OF NAVIGATION, 2010, 63 (01): : 1 - 22
  • [8] A New Tightly-Coupled Dipole Array
    Tong, San-Qiang
    Wang, Bing-Zhong
    Wang, Ren
    Hu, Chang-Hai
    2020 IEEE INTERNATIONAL SYMPOSIUM ON ANTENNAS AND PROPAGATION AND NORTH AMERICAN RADIO SCIENCE MEETING, 2020, : 139 - 140
  • [9] Reconfiguration in tightly-coupled conferencing environments
    Trossen, D
    Eickhoff, WC
    CHALLENGES OF INFORMATION TECHNOLOGY MANAGEMENT IN THE 21ST CENTURY, 2000, : 156 - 160
  • [10] Tightly-coupled Convolutional Neural Network with Spatial-temporal Memory for Text Classification
    Wang, Shiyao
    Deng, Zhidong
    2017 INTERNATIONAL JOINT CONFERENCE ON NEURAL NETWORKS (IJCNN), 2017, : 2370 - 2376