A Transformer-Based Hierarchical Variational AutoEncoder Combined Hidden Markov Model for Long Text Generation

被引:8
|
作者
Zhao, Kun [1 ]
Ding, Hongwei [1 ]
Ye, Kai [1 ]
Cui, Xiaohui [1 ]
机构
[1] Wuhan Univ, Sch Cyber Sci & Engn, Key Lab Aerosp Informat Secur & Trusted Comp, Minist Educ, Wuhan 430072, Peoples R China
关键词
Variational AutoEncoder; text generation; Hidden Markov Model; Transformer; latent variables;
D O I
10.3390/e23101277
中图分类号
O4 [物理学];
学科分类号
0702 ;
摘要
The Variational AutoEncoder (VAE) has made significant progress in text generation, but it focused on short text (always a sentence). Long texts consist of multiple sentences. There is a particular relationship between each sentence, especially between the latent variables that control the generation of the sentences. The relationships between these latent variables help in generating continuous and logically connected long texts. There exist very few studies on the relationships between these latent variables. We proposed a method for combining the Transformer-Based Hierarchical Variational AutoEncoder and Hidden Markov Model (HT-HVAE) to learn multiple hierarchical latent variables and their relationships. This application improves long text generation. We use a hierarchical Transformer encoder to encode the long texts in order to obtain better hierarchical information of the long text. HT-HVAE's generation network uses HMM to learn the relationship between latent variables. We also proposed a method for calculating the perplexity for the multiple hierarchical latent variable structure. The experimental results show that our model is more effective in the dataset with strong logic, alleviates the notorious posterior collapse problem, and generates more continuous and logically connected long text.
引用
收藏
页数:18
相关论文
共 50 条
  • [21] De Novo Generation of Chemical Structures of Inhibitor and Activator Candidates for Therapeutic Target Proteins by a Transformer-Based Variational Autoencoder and Bayesian Optimization
    Matsukiyo, Yuki
    Yamanaka, Chikashige
    Yamanishi, Yoshihiro
    JOURNAL OF CHEMICAL INFORMATION AND MODELING, 2023, 64 (07) : 2345 - 2355
  • [22] LayoutDM: Transformer-based Diffusion Model for Layout Generation
    Chai, Shang
    Zhuang, Liansheng
    Yan, Fengying
    2023 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR), 2023, : 18349 - 18358
  • [23] An Adversarial Hierarchical Hidden Markov Model for Human Pose Modeling and Generation
    Zhao, Rui
    Ji, Qiang
    THIRTY-SECOND AAAI CONFERENCE ON ARTIFICIAL INTELLIGENCE / THIRTIETH INNOVATIVE APPLICATIONS OF ARTIFICIAL INTELLIGENCE CONFERENCE / EIGHTH AAAI SYMPOSIUM ON EDUCATIONAL ADVANCES IN ARTIFICIAL INTELLIGENCE, 2018, : 2636 - 2643
  • [24] Anomaly detection in KOMAC high-power systems using transformer-based conditional variational autoencoder
    Kim, Gi-Hu
    Jeong, Hae-Seong
    Kim, Han-Sung
    Kwon, Hyeok-Jung
    Kim, Dong-Hwan
    JOURNAL OF THE KOREAN PHYSICAL SOCIETY, 2025,
  • [25] Investigating African-American Vernacular English in Transformer-Based Text Generation
    Groenwold, Sophie
    Ou, Lily
    Parekh, Aesha
    Honnavalli, Samhita
    Levy, Sharon
    Mirza, Diba
    Wang, William Yang
    PROCEEDINGS OF THE 2020 CONFERENCE ON EMPIRICAL METHODS IN NATURAL LANGUAGE PROCESSING (EMNLP), 2020, : 5877 - 5883
  • [26] TILGAN: Transformer-based Implicit Latent GAN for Diverse and Coherent Text Generation
    Diao, Shizhe
    Shen, Xinwei
    Shum, KaShun
    Song, Yan
    Zhang, Tong
    FINDINGS OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS, ACL-IJCNLP 2021, 2021, : 4844 - 4858
  • [27] Neural Rule-Execution Tracking Machine For Transformer-Based Text Generation
    Wang, Yufei
    Xu, Can
    Hu, Huang
    Tao, Chongyang
    Wan, Stephen
    Dras, Mark
    Johnson, Mark
    Jiang, Daxin
    ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 34 (NEURIPS 2021), 2021, 34
  • [28] A Transformer-Based Model for Multi-Track Music Generation
    Jin, Cong
    Wang, Tao
    Liu, Shouxun
    Tie, Yun
    Li, Jianguang
    Li, Xiaobing
    Lui, Simon
    INTERNATIONAL JOURNAL OF MULTIMEDIA DATA ENGINEERING & MANAGEMENT, 2020, 11 (03): : 36 - 54
  • [29] Hierarchical Graph Transformer-Based Deep Learning Model for Large-Scale Multi-Label Text Classification
    Gong, Jibing
    Teng, Zhiyong
    Teng, Qi
    Zhang, Hekai
    Du, Linfeng
    Chen, Shuai
    Bhuiyan, Md Zakirul Alam
    Li, Jianhua
    Liu, Mingsheng
    Ma, Hongyuan
    IEEE ACCESS, 2020, 8 : 30885 - 30896
  • [30] Mechanical Fault Diagnosis Based on Variational Bayesian Hidden Markov Model
    Zhang, Xiqin
    Li, Zhinong
    Tu, Jingzhi
    2018 PROGNOSTICS AND SYSTEM HEALTH MANAGEMENT CONFERENCE (PHM-CHONGQING 2018), 2018, : 1250 - 1254