Training Heterogeneous Features in Sequence to Sequence Tasks: Latent Enhanced Multi-filter Seq2Seq Model

被引:0
|
作者
Yang, Yunhao [1 ]
Xue, Zhaokun [1 ]
机构
[1] Univ Texas Austin, Austin, TX 78705 USA
关键词
Representation learning; Recurrent neural network; Latent subspace; Neural machine translation; Semantic parsing;
D O I
10.1007/978-3-031-16078-3_6
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
In language processing, training data with extremely large variance may lead to difficulty of language model's convergence. It is difficult for the network parameters to adapt sentences with largely varied semantics or grammatical structures. To resolve this problem, we introduce a model that concentrates the each of the heterogeneous features in the input sentences. Build upon the encoder-decoder architecture, we design a latent-enhanced multi-filter seq2seq model (LEMS) that analyzes the input representations by introducing a latent space transformation and clustering. The representations are extracted from the final hidden state of the encoder and lie in the latent space. A latent space transformation is applied for enhancing the quality of the representations. Thus the clustering algorithm can easily separate samples based on the features of these representations. Multiple filters are trained by the features from their corresponding clusters, the heterogeneity of the training data can be resolved accordingly. We conduct two sets of comparative experiments on semantic parsing and machine translation, using the Geo-query dataset and Multi30k English-French to demonstrate the enhancement our model has made, respectively.
引用
收藏
页码:103 / 117
页数:15
相关论文
共 29 条
  • [1] Survival Seq2Seq: A Survival Model based on Sequence to Sequence Architecture
    Pourjafari, Ebrahim
    Ziaei, Navid
    Rezaei, Mohammad R.
    Sameizadeh, Amir
    Shafiee, Mohammad
    Alavinia, Mohammad
    Abolghasemian, Mansour
    Sajadi, Nick
    MACHINE LEARNING FOR HEALTHCARE CONFERENCE, VOL 182, 2022, 182 : 79 - 100
  • [2] Survival Seq2Seq: A Survival Model based on Sequence to Sequence Architecture
    Pourjafari, Ebrahim
    Ziaei, Navid
    Rezaei, Mohammad R.
    Sameizadeh, Amir
    Shafiee, Mohammad
    Alavinia, Mohammad
    Abolghasemian, Mansour
    Sajadi, Nick
    Proceedings of Machine Learning Research, 2022, 182 : 79 - 100
  • [3] Adaptive Multistep Prediction With Sequence-to-Sequence (Seq2Seq) Models
    Kelley, Joseph
    Hagan, Martin
    IEEE TRANSACTIONS ON NEURAL NETWORKS AND LEARNING SYSTEMS, 2025,
  • [4] A Seq2Seq Model Improved by Transcendental Learning and Imaged Sequence Samples for Porosity Prediction
    Zhou, Lijian
    Wang, Lijun
    Zhao, Zhiang
    Liu, Yuwei
    Liu, Xiwu
    MATHEMATICS, 2023, 11 (01)
  • [5] Predicting MicroRNA Sequence Using CNN and LSTM Stacked in Seq2Seq Architecture
    Chakraborty, Rajkumar
    Hasija, Yasha
    IEEE-ACM TRANSACTIONS ON COMPUTATIONAL BIOLOGY AND BIOINFORMATICS, 2020, 17 (06) : 2183 - 2188
  • [6] Abstractive Summarization Model with a Feature-Enhanced Seq2Seq Structure
    Hao, Zepeng
    Ji, Jingzhou
    Xie, Tao
    Xue, Bin
    2020 5TH ASIA-PACIFIC CONFERENCE ON INTELLIGENT ROBOT SYSTEMS (ACIRS 2020), 2020, : 163 - 167
  • [7] Seq2Seq-AFL: Fuzzing via sequence-to-sequence model
    Yang, Liqun
    Wei, Chaoren
    Yang, Jian
    Ma, Jinxin
    Guo, Hongcheng
    Cheng, Long
    Li, Zhoujun
    INTERNATIONAL JOURNAL OF MACHINE LEARNING AND CYBERNETICS, 2024, 15 (10) : 4403 - 4421
  • [8] GSSF: A Generative Sequence Similarity Function Based on a Seq2Seq Model for Clustering Online Handwritten Mathematical Answers
    Huy Quang Ung
    Cuong Tuan Nguyen
    Hung Tuan Nguyen
    Nakagawa, Masaki
    DOCUMENT ANALYSIS AND RECOGNITION - ICDAR 2021, PT II, 2021, 12822 : 145 - 159
  • [9] Viewing the MIMO Channel as Sequence Rather than Image: A Seq2Seq Approach for Efficient CSI Feedback
    Chen, Zirui
    Zhang, Zhaoyang
    Xiao, Zhuoran
    2022 IEEE WIRELESS COMMUNICATIONS AND NETWORKING CONFERENCE (WCNC), 2022, : 2292 - 2297
  • [10] Multi-Label Classification of Options Based on Seq2seq Model of Hybrid Attention
    Chen, Qian
    Han, Lin
    Wang, Suge
    Guo, Xin
    Computer Engineering and Applications, 2024, 59 (04) : 104 - 111