A Transformer-Based Hierarchical Variational AutoEncoder Combined Hidden Markov Model for Long Text Generation

被引:8
|
作者
Zhao, Kun [1 ]
Ding, Hongwei [1 ]
Ye, Kai [1 ]
Cui, Xiaohui [1 ]
机构
[1] Wuhan Univ, Sch Cyber Sci & Engn, Key Lab Aerosp Informat Secur & Trusted Comp, Minist Educ, Wuhan 430072, Peoples R China
关键词
Variational AutoEncoder; text generation; Hidden Markov Model; Transformer; latent variables;
D O I
10.3390/e23101277
中图分类号
O4 [物理学];
学科分类号
0702 ;
摘要
The Variational AutoEncoder (VAE) has made significant progress in text generation, but it focused on short text (always a sentence). Long texts consist of multiple sentences. There is a particular relationship between each sentence, especially between the latent variables that control the generation of the sentences. The relationships between these latent variables help in generating continuous and logically connected long texts. There exist very few studies on the relationships between these latent variables. We proposed a method for combining the Transformer-Based Hierarchical Variational AutoEncoder and Hidden Markov Model (HT-HVAE) to learn multiple hierarchical latent variables and their relationships. This application improves long text generation. We use a hierarchical Transformer encoder to encode the long texts in order to obtain better hierarchical information of the long text. HT-HVAE's generation network uses HMM to learn the relationship between latent variables. We also proposed a method for calculating the perplexity for the multiple hierarchical latent variable structure. The experimental results show that our model is more effective in the dataset with strong logic, alleviates the notorious posterior collapse problem, and generates more continuous and logically connected long text.
引用
收藏
页数:18
相关论文
共 50 条
  • [31] HRPE: Hierarchical Relative Positional Encoding for Transformer-Based Structured Symbolic Music Generation
    Li, Pengfei
    Wu, Jingcheng
    Ji, Zihao
    MUSIC INTELLIGENCE, SOMI 2023, 2024, 2007 : 122 - 134
  • [32] PETR: Rethinking the Capability of Transformer-Based Language Model in Scene Text Recognition
    Wang, Yuxin
    Xie, Hongtao
    Fang, Shancheng
    Xing, Mengting
    Wang, Jing
    Zhu, Shenggao
    Zhang, Yongdong
    IEEE TRANSACTIONS ON IMAGE PROCESSING, 2022, 31 : 5585 - 5598
  • [33] Text Information Extraction based on Genetic Algorithm and Hidden Markov Model
    Li, Rong
    Zheng, Jia-heng
    Pei, Chun-qin
    PROCEEDINGS OF THE FIRST INTERNATIONAL WORKSHOP ON EDUCATION TECHNOLOGY AND COMPUTER SCIENCE, VOL I, 2009, : 334 - +
  • [34] A hidden Markov model-based text classification of medical documents
    Yi, Kwan
    Beheshti, Jamshid
    JOURNAL OF INFORMATION SCIENCE, 2009, 35 (01) : 67 - 81
  • [35] Gait Recognition Based on GFHI and Combined Hidden Markov Model
    Chen, Kai
    Wu, Shiyu
    Li, Zhihua
    2020 13TH INTERNATIONAL CONGRESS ON IMAGE AND SIGNAL PROCESSING, BIOMEDICAL ENGINEERING AND INFORMATICS (CISP-BMEI 2020), 2020, : 287 - 292
  • [36] A novel transformer-based graph generation model for vectorized road design
    Zhou, Peichi
    Li, Chen
    Zhang, Jian
    Wang, Changbo
    Qin, Hong
    Liu, Long
    COMPUTER ANIMATION AND VIRTUAL WORLDS, 2024, 35 (03)
  • [37] BAYESIAN LINEAR REGRESSION FOR HIDDEN MARKOV MODEL BASED ON OPTIMIZING VARIATIONAL BOUNDS
    Watanabe, Shinji
    Nakamura, Atsushi
    Luang, Biing-Hwang
    2011 IEEE INTERNATIONAL WORKSHOP ON MACHINE LEARNING FOR SIGNAL PROCESSING (MLSP), 2011,
  • [38] Chinese named entity recognition method based on Transformer and hidden Markov model
    Li J.
    Xiong Q.
    Hu Y.-T.
    Liu K.-Y.
    Jilin Daxue Xuebao (Gongxueban)/Journal of Jilin University (Engineering and Technology Edition), 2023, 53 (05): : 1427 - 1434
  • [39] A Survey of Controllable Text Generation Using Transformer-based Pre-trained Language Models
    Zhang, Hanqing
    Song, Haolin
    Li, Shaoyu
    Zhou, Ming
    Song, Dawei
    ACM COMPUTING SURVEYS, 2024, 56 (03)
  • [40] Anomaly Detection in Mutual Actions: Unsupervised Classification of Fighting and Non-Fighting Behaviors Using Transformer-Based Variational Autoencoder
    Zaw, Thura
    Komuro, Takashi
    ADVANCES IN VISUAL COMPUTING, ISVC 2024, PT I, 2025, 15046 : 397 - 410