Mixture Variational Autoencoder of Boltzmann Machines for Text Processing

被引:0
|
作者
Guilherme Gomes, Bruno [1 ]
Murai, Fabricio [1 ]
Goussevskaia, Olga [1 ]
Couto Da Silva, Ana Paula [1 ]
机构
[1] Univ Fed Minas Gerais, Belo Horizonte, MG, Brazil
关键词
D O I
10.1007/978-3-030-80599-9_5
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Variational autoencoders (VAEs) have been successfully used to learn good representations in unsupervised settings, especially for image data. More recently, mixture variational autoencoders (MVAEs) have been proposed to enhance the representation capabilities of VAEs by assuming that data can come from a mixture distribution. In this work, we adapt MVAEs for text processing by modeling each component's joint distribution of latent variables and document's bag-of-words as a graphical model known as the Boltzmann Machine, popular in natural language processing for performing well in a number of tasks. The proposed model, MVAE-BM, can learn text representations from unlabeled data without requiring pre-trained word embeddings. We evaluate the representations obtained by MVAE-BM on six corpora w.r.t. the perplexity metric and accuracy on binary and multi-class text classification. Despite its simplicity, our results show that MVAE-BM's performance is on par with or superior to that of modern deep learning techniques such as BERT and RoBERTa. Last, we show that the mapping to mixture components learned by the model lends itself naturally to document clustering.
引用
收藏
页码:46 / 56
页数:11
相关论文
共 50 条
  • [41] Variational Selective Autoencoder
    Gong, Yu
    Hajimirsadeghi, Hossein
    He, Jiawei
    Nawhal, Megha
    Durand, Thibaut
    Mori, Greg
    SYMPOSIUM ON ADVANCES IN APPROXIMATE BAYESIAN INFERENCE, VOL 118, 2019, 118
  • [42] Longitudinal Variational Autoencoder
    Ramchandran, Siddharth
    Tikhonov, Gleb
    Kujanpaa, Kalle
    Koskinen, Miika
    Lahdesmaki, Harri
    24TH INTERNATIONAL CONFERENCE ON ARTIFICIAL INTELLIGENCE AND STATISTICS (AISTATS), 2021, 130
  • [43] The Variational InfoMax AutoEncoder
    Crescimanna, Vincenzo
    Graham, Bruce
    2020 INTERNATIONAL JOINT CONFERENCE ON NEURAL NETWORKS (IJCNN), 2020,
  • [44] Generative autoencoder to prevent overregularization of variational autoencoder
    Ko, YoungMin
    Ko, SunWoo
    Kim, YoungSoo
    ETRI JOURNAL, 2025, 47 (01) : 80 - 89
  • [45] Mixture autoregressive and spectral attention network for multispectral image compression based on variational autoencoder
    Kong, Fanqiang
    Ren, Guanglong
    Hu, Yunfang
    Li, Dan
    Hu, Kedi
    VISUAL COMPUTER, 2024, 40 (09): : 6295 - 6318
  • [46] A Novel Model for Ship Trajectory Anomaly Detection Based on Gaussian Mixture Variational Autoencoder
    Xie, Lei
    Guo, Tao
    Chang, Jiliang
    Wan, Chengpeng
    Hu, Xinyuan
    Yang, Yang
    Ou, Changkui
    IEEE TRANSACTIONS ON VEHICULAR TECHNOLOGY, 2023, 72 (11) : 13826 - 13835
  • [47] Spoken-Text-Style Transfer with Conditional Variational Autoencoder and Content Word Storage
    Yoshioka, Daiki
    Yasuda, Yusuke
    Matsunaga, Noriyuki
    Ohtani, Yamato
    Toda, Tomoki
    INTERSPEECH 2022, 2022, : 4576 - 4580
  • [48] Enhancing Graph Variational Autoencoder for Short Text Topic Modeling with Mutual Information Maximization
    Ge, Yuhang
    Hu, Xuegang
    2022 IEEE INTERNATIONAL CONFERENCE ON KNOWLEDGE GRAPH (ICKG), 2022, : 64 - 70
  • [49] Video anomaly detection and localization via Gaussian Mixture Fully Convolutional Variational Autoencoder
    Fan, Yaxiang
    Wen, Gongjian
    Li, Deren
    Qiu, Shaohua
    Levine, Martin D.
    Xiao, Fei
    COMPUTER VISION AND IMAGE UNDERSTANDING, 2020, 195
  • [50] A GAUSSIAN MIXTURE VARIATIONAL AUTOENCODER-BASED APPROACH FOR DESIGNING PHONONIC BANDGAP METAMATERIALS
    Wang, Zihan
    Xian, Weikang
    Baccouche, M. Ridha
    Lanzerath, Horst
    Li, Ying
    Xu, Hongyi
    PROCEEDINGS OF ASME 2021 INTERNATIONAL DESIGN ENGINEERING TECHNICAL CONFERENCES AND COMPUTERS AND INFORMATION IN ENGINEERING CONFERENCE, IDETC-CIE2021, VOL 3B, 2021,