Improving stock trend prediction with pretrain multi-granularity denoising contrastive learning

被引:0
|
作者
Wang, Mingjie [2 ]
Wang, Siyuan [3 ]
Guo, Jianxiong [1 ,2 ]
Jia, Weijia [1 ,2 ]
机构
[1] Beijing Normal Univ, Adv Inst Nat Sci, Zhuhai 519087, Peoples R China
[2] BNU, HKBU United Int Coll, Dept Comp Sci, Guangdong Key Lab AI & Multimodal Data Proc, Zhuhai 519087, Peoples R China
[3] BNU, Fac Sci & Technol, Dept Math Sci, HKBU United Int Coll, Zhuhai 519087, Peoples R China
基金
中国国家自然科学基金;
关键词
Contrastive learning; Multi-granularity data; Memory; Denoising; Stock trend prediction; Pre-training; TIME-SERIES; NOISY DATA; CLASSIFICATION; INDEX; MODEL;
D O I
10.1007/s10115-023-02006-1
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Stock trend prediction (STP) aims to predict price fluctuation, which is critical in financial trading. The existing STP approaches only use market data with the same granularity (e.g., as daily market data). However, in the actual financial investment, there are a large number of more detailed investment signals contained in finer-grained data (e.g., high-frequency data). This motivates us to research how to leverage multi-granularity market data to capture more useful information and improve the accuracy in the task of STP. However, the effective utilization of multi-granularity data presents a major challenge. Firstly, the iteration of multi-granularity data with time will lead to more complex noise, which is difficult to extract signals. Secondly, the difference in granularity may lead to opposite target trends in the same time interval. Thirdly, the target trends of stocks with similar features can be quite different, and different sizes of granularity will aggravate this gap. In order to address these challenges, we present a self-supervised framework of multi-granularity denoising contrastive learning (MDC). Specifically, we construct a dynamic dictionary of memory, which can obtain clear and unified representations by filtering noise and aligning multi-granularity data. Moreover, we design two contrast learning modules during the fine-tuning stage to solve the differences in trends by constructing additional self-supervised signals. Besides, in the pre-training stage, we design the granularity domain adaptation module (GDA) to address the issues of temporal inconsistency and data imbalance associated with different granularity in financial data, alongside the memory self-distillation module (MSD) to tackle the challenge posed by a low signal-to-noise ratio. The GDA alleviates these complications by replacing a portion of the coarse-grained data with the preceding time step's fine-grained data, while the MSD seeks to filter out intrinsic noise by aligning the fine-grained representations with the coarse-grained representations' distribution using a self-distillation mechanism. Extensive experiments on the CSI 300 and CSI 100 datasets show that our framework stands out from the existing top-level systems and has excellent profitability in real investing scenarios.
引用
收藏
页码:2439 / 2466
页数:28
相关论文
共 50 条
  • [41] Multi-granularity feature learning network for deep hashing
    Feng, Hao
    Wang, Nian
    Tang, Jun
    Chen, Jie
    Chen, Feng
    NEUROCOMPUTING, 2021, 423 : 274 - 283
  • [42] A review on network representation learning with multi-granularity perspective
    Fu, Shun
    Wang, Lufeng
    Yang, Jie
    INTELLIGENT DATA ANALYSIS, 2024, 28 (01) : 3 - 32
  • [43] Multi-granularity network representation learning on overlapping communities
    Zhou, Rongrong
    Li, Jinhai
    INTERNATIONAL JOURNAL OF MACHINE LEARNING AND CYBERNETICS, 2024, 15 (07) : 2935 - 2955
  • [44] Text-Based Occluded Person Re-identification via Multi-Granularity Contrastive Consistency Learning
    Wu, Xinyi
    Ma, Wentao
    Guo, Dan
    Zhou, Tongqing
    Zhao, Shan
    Cai, Zhiping
    THIRTY-EIGHTH AAAI CONFERENCE ON ARTIFICIAL INTELLIGENCE, VOL 38 NO 6, 2024, : 6162 - 6170
  • [45] MGCNet: Multi-granularity cataract classification using denoising diffusion probabilistic model ☆
    Wang, Qiong
    Sun, Hongdi
    Feng, Yu
    Dong, Zhe
    Bai, Cong
    DISPLAYS, 2024, 83
  • [46] Multi-granularity Semantic and Acoustic Stress Prediction for Expressive TTS
    Chi, Wenjiang
    Feng, Xiaoqin
    Xue, Liumeng
    Chen, Yunlin
    Xie, Lei
    Li, Zhifei
    2023 ASIA PACIFIC SIGNAL AND INFORMATION PROCESSING ASSOCIATION ANNUAL SUMMIT AND CONFERENCE, APSIPA ASC, 2023, : 2409 - 2415
  • [47] M2Mol: Multi-view Multi-granularity Molecular Representation Learning for Property Prediction
    Zhang, Ran
    Wang, Xuezhi
    Liu, Kunpeng
    Zhou, Yuanchun
    Wang, Pengfei
    DATABASE SYSTEMS FOR ADVANCED APPLICATIONS, PT VII, DASFAA 2024, 2024, 14856 : 264 - 274
  • [48] MGFL: Multi-granularity Federated Learning in Edge Computing Systems
    Cai, Shangxuan
    Zhao, Yunfeng
    Liu, Zhicheng
    Qiu, Chao
    Wang, Xiaofei
    Hu, Qinghua
    ALGORITHMS AND ARCHITECTURES FOR PARALLEL PROCESSING, ICA3PP 2021, PT I, 2022, 13155 : 549 - 563
  • [49] Multi-Granularity Anchor-Contrastive Representation Learning for Semi-Supervised Skeleton-Based Action Recognition
    Shu, Xiangbo
    Xu, Binqian
    Zhang, Liyan
    Tang, Jinhui
    IEEE TRANSACTIONS ON PATTERN ANALYSIS AND MACHINE INTELLIGENCE, 2023, 45 (06) : 7559 - 7576
  • [50] Multi-Granularity Weighted Federated Learning for Heterogeneous Edge Computing
    Zhao, Yunfeng
    Qiu, Chao
    Cai, Shangxuan
    Liu, Zhicheng
    Wang, Yu
    Wang, Xiaofei
    Hu, Qinghua
    IEEE TRANSACTIONS ON SERVICES COMPUTING, 2025, 18 (01) : 270 - 287