A Transformer-Based Hierarchical Variational AutoEncoder Combined Hidden Markov Model for Long Text Generation

被引:8
|
作者
Zhao, Kun [1 ]
Ding, Hongwei [1 ]
Ye, Kai [1 ]
Cui, Xiaohui [1 ]
机构
[1] Wuhan Univ, Sch Cyber Sci & Engn, Key Lab Aerosp Informat Secur & Trusted Comp, Minist Educ, Wuhan 430072, Peoples R China
关键词
Variational AutoEncoder; text generation; Hidden Markov Model; Transformer; latent variables;
D O I
10.3390/e23101277
中图分类号
O4 [物理学];
学科分类号
0702 ;
摘要
The Variational AutoEncoder (VAE) has made significant progress in text generation, but it focused on short text (always a sentence). Long texts consist of multiple sentences. There is a particular relationship between each sentence, especially between the latent variables that control the generation of the sentences. The relationships between these latent variables help in generating continuous and logically connected long texts. There exist very few studies on the relationships between these latent variables. We proposed a method for combining the Transformer-Based Hierarchical Variational AutoEncoder and Hidden Markov Model (HT-HVAE) to learn multiple hierarchical latent variables and their relationships. This application improves long text generation. We use a hierarchical Transformer encoder to encode the long texts in order to obtain better hierarchical information of the long text. HT-HVAE's generation network uses HMM to learn the relationship between latent variables. We also proposed a method for calculating the perplexity for the multiple hierarchical latent variable structure. The experimental results show that our model is more effective in the dataset with strong logic, alleviates the notorious posterior collapse problem, and generates more continuous and logically connected long text.
引用
收藏
页数:18
相关论文
共 50 条
  • [1] A Transformer-Based Variational Autoencoder for Sentence Generation
    Liu, Danyang
    Liu, Gongshen
    2019 INTERNATIONAL JOINT CONFERENCE ON NEURAL NETWORKS (IJCNN), 2019,
  • [2] Transformer-Based Direct Hidden Markov Model for Machine Translation
    Wang, Weiyue
    Yang, Zijian
    Gao, Yingbo
    Ney, Hermann
    ACL-IJCNLP 2021: THE 59TH ANNUAL MEETING OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS AND THE 11TH INTERNATIONAL JOINT CONFERENCE ON NATURAL LANGUAGE PROCESSING: PROCEEDINGS OF THE STUDENT RESEARCH WORKSHOP, 2021, : 23 - 32
  • [3] Long and Diverse Text Generation with Planning-based Hierarchical Variational Model
    Shao, Zhihong
    Huang, Minlie
    Wen, Jiangtao
    Xu, Wenfei
    Zhu, Xiaoyan
    2019 CONFERENCE ON EMPIRICAL METHODS IN NATURAL LANGUAGE PROCESSING AND THE 9TH INTERNATIONAL JOINT CONFERENCE ON NATURAL LANGUAGE PROCESSING (EMNLP-IJCNLP 2019): PROCEEDINGS OF THE CONFERENCE, 2019, : 3257 - 3268
  • [4] Hidden Markov Model Variational Autoencoder for Acoustic Unit Discovery
    Ebbers, Janek
    Heymann, Jahn
    Drude, Lukas
    Glarner, Thomas
    Haeb-Umbach, Reinhold
    Raj, Bhiksha
    18TH ANNUAL CONFERENCE OF THE INTERNATIONAL SPEECH COMMUNICATION ASSOCIATION (INTERSPEECH 2017), VOLS 1-6: SITUATED INTERACTION, 2017, : 488 - 492
  • [5] T-DVAE: A Transformer-Based Dynamical Variational Autoencoder for Speech
    Perschewski, Jan-Ole
    Stober, Sebastian
    ARTIFICIAL NEURAL NETWORKS AND MACHINE LEARNING-ICANN 2024, PT VII, 2024, 15022 : 33 - 46
  • [6] Adaptive Transformer-Based Conditioned Variational Autoencoder for Incomplete Social Event Classification
    Li, Zhangming
    Qian, Shengsheng
    Cao, Jie
    Fang, Quan
    Xu, Changsheng
    PROCEEDINGS OF THE 30TH ACM INTERNATIONAL CONFERENCE ON MULTIMEDIA, MM 2022, 2022, : 1698 - 1707
  • [7] T-CVAE: Transformer-Based Conditioned Variational Autoencoder for Story Completion
    Wang, Tianming
    Wan, Xiaojun
    PROCEEDINGS OF THE TWENTY-EIGHTH INTERNATIONAL JOINT CONFERENCE ON ARTIFICIAL INTELLIGENCE, 2019, : 5233 - 5239
  • [8] Transformer-based Question Text Generation in the Learning System
    Li, Jiajun
    Song, Huazhu
    Li, Jun
    6TH INTERNATIONAL CONFERENCE ON INNOVATION IN ARTIFICIAL INTELLIGENCE, ICIAI2022, 2022, : 50 - 56
  • [9] Applying Transformer-Based Text Summarization for Keyphrase Generation
    Glazkova A.V.
    Morozov D.A.
    Lobachevskii Journal of Mathematics, 2023, 44 (1) : 123 - 136
  • [10] A transformer-based approach to Nigerian Pidgin text generation
    Garba, Kabir
    Kolajo, Taiwo
    Agbogun, Joshua B.
    International Journal of Speech Technology, 2024, 27 (04) : 1027 - 1037