Mixture Variational Autoencoder of Boltzmann Machines for Text Processing

被引:0
|
作者
Guilherme Gomes, Bruno [1 ]
Murai, Fabricio [1 ]
Goussevskaia, Olga [1 ]
Couto Da Silva, Ana Paula [1 ]
机构
[1] Univ Fed Minas Gerais, Belo Horizonte, MG, Brazil
关键词
D O I
10.1007/978-3-030-80599-9_5
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Variational autoencoders (VAEs) have been successfully used to learn good representations in unsupervised settings, especially for image data. More recently, mixture variational autoencoders (MVAEs) have been proposed to enhance the representation capabilities of VAEs by assuming that data can come from a mixture distribution. In this work, we adapt MVAEs for text processing by modeling each component's joint distribution of latent variables and document's bag-of-words as a graphical model known as the Boltzmann Machine, popular in natural language processing for performing well in a number of tasks. The proposed model, MVAE-BM, can learn text representations from unlabeled data without requiring pre-trained word embeddings. We evaluate the representations obtained by MVAE-BM on six corpora w.r.t. the perplexity metric and accuracy on binary and multi-class text classification. Despite its simplicity, our results show that MVAE-BM's performance is on par with or superior to that of modern deep learning techniques such as BERT and RoBERTa. Last, we show that the mapping to mixture components learned by the model lends itself naturally to document clustering.
引用
收藏
页码:46 / 56
页数:11
相关论文
共 50 条
  • [31] A Unified Unsupervised Gaussian Mixture Variational Autoencoder for High Dimensional Outlier Detection
    Liao, Weixian
    Guo, Yifan
    Chen, Xuhui
    Li, Pan
    2018 IEEE INTERNATIONAL CONFERENCE ON BIG DATA (BIG DATA), 2018, : 1208 - 1217
  • [32] Condition-Transforming Variational Autoencoder for Generating Diverse Short Text Conversations
    Ruan, Yu-Ping
    Ling, Zhen-Hua
    Zhu, Xiaodan
    ACM TRANSACTIONS ON ASIAN AND LOW-RESOURCE LANGUAGE INFORMATION PROCESSING, 2020, 19 (06)
  • [33] An Improved Semi-supervised Variational Autoencoder with Gate Mechanism for Text Classification
    Ye, Haiming
    Zhang, Weiwen
    Nie, Mengna
    INTERNATIONAL JOURNAL OF PATTERN RECOGNITION AND ARTIFICIAL INTELLIGENCE, 2022, 36 (10)
  • [34] Condition-Transforming Variational Autoencoder for Generating Diverse Short Text Conversations
    Ruan, Yu-Ping
    Ling, Zhen-Hua
    Zhu, Xiaodan
    Ruan, Yu-Ping (ypruan@mail.ustc.edu.cn); Ling, Zhen-Hua (zhling@ustc.edu.cn), 1600, Association for Computing Machinery (19):
  • [35] Text-mining the NeuroSynth corpus using Deep Boltzmann Machines
    Monti, Ricardo
    Lorenz, Romy
    Leech, Robert
    Anagnostopoulos, Christoforos
    Montana, Giovanni
    2016 6TH INTERNATIONAL WORKSHOP ON PATTERN RECOGNITION IN NEUROIMAGING (PRNI), 2016, : 13 - 16
  • [36] Unsupervised Clustering through Gaussian Mixture Variational AutoEncoder with Non-Reparameterized Variational Inference and Std Annealing
    Li, Zhihan
    Zhao, Youjian
    Xu, Haowen
    Chen, Wenxiao
    Xu, Shangqing
    Li, Yilin
    Pei, Dan
    2020 INTERNATIONAL JOINT CONFERENCE ON NEURAL NETWORKS (IJCNN), 2020,
  • [37] The Autoencoding Variational Autoencoder
    Cemgil, A. Taylan
    Ghaisas, Sumedh
    Dvijotham, Krishnamurthy
    Gowal, Sven
    Kohli, Pushmeet
    ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 33, NEURIPS 2020, 2020, 33
  • [38] Dirichlet Variational Autoencoder
    Joo, Weonyoung
    Lee, Wonsung
    Park, Sungrae
    Moon, Il-Chul
    PATTERN RECOGNITION, 2020, 107
  • [39] Grammar Variational Autoencoder
    Kusner, Matt J.
    Paige, Brooks
    Hernandez-Lobato, Jose Miguel
    INTERNATIONAL CONFERENCE ON MACHINE LEARNING, VOL 70, 2017, 70
  • [40] Quantum variational autoencoder
    Khoshaman, Amir
    Vinci, Walter
    Denis, Brandon
    Andriyash, Evgeny
    Amin, Mohammad H.
    QUANTUM SCIENCE AND TECHNOLOGY, 2019, 4 (01)