Discrete Auto-regressive Variational Attention Models for Text Modeling

被引:0
|
作者
Fang, Xianghong [1 ]
Bai, Haoli [1 ]
Li, Jian [1 ]
Xu, Zenglin [2 ]
Lyu, Michael [1 ]
King, Irwin [1 ]
机构
[1] Chinese Univ Hong Kong, Dept Comp Sci & Engn, Hong Kong, Peoples R China
[2] Harbin Inst Technol, Sch Comp Sci & Engn, Shenzhen, Peoples R China
关键词
Text Modeling; Information Underrepresentation; Posterior Collapse;
D O I
10.1109/IJCNN52387.2021.9534375
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Variational autoencoders (VAEs) have been widely applied for text modeling. In practice, however, they are troubled by two challenges: information underrepresentation and posterior collapse. The former arises as only the last hidden state of LSTM encoder is transformed into the latent space, which is generally insufficient to summarize the data. The latter is a long-standing problem during the training of VAEs as the optimization is trapped to a disastrous local optimum. In this paper, we propose Discrete Auto-regressive Variational Attention Model (DAVAM) to address the challenges. Specifically, we introduce an auto-regressive variational attention approach to enrich the latent space by effectively capturing the semantic dependency from the input. We further design discrete latent space for the variational attention and mathematically show that our model is free from posterior collapse. Extensive experiments on language modeling tasks demonstrate the superiority of DAVAM against several VAE counterparts. Code will be released.
引用
收藏
页数:8
相关论文
共 50 条
  • [31] Auto-Regressive Discrete Acquisition Points Transformation for Diffusion Weighted MRI Data
    Metcalfe-Smith, Emma
    Meeus, Emma M.
    Novak, Jan
    Dehghani, Hamid
    Peet, Andrew C.
    Zarinabad, Niloufar
    IEEE TRANSACTIONS ON BIOMEDICAL ENGINEERING, 2019, 66 (09) : 2617 - 2628
  • [32] Penalized estimation of threshold auto-regressive models with many components and thresholds
    Zhang, Kunhui
    Safikhani, Abolfazl
    Tank, Alex
    Shojaie, Ali
    ELECTRONIC JOURNAL OF STATISTICS, 2022, 16 (01): : 1891 - 1951
  • [33] Correcting Multivariate Auto-Regressive Models for the Influence of Unobserved Common Input
    Gomez, Vicenc
    Gheshlaghi Azar, Mohammad
    Kappen, Hilbert J.
    ARTIFICIAL INTELLIGENCE RESEARCH AND DEVELOPMENT, 2016, 288 : 177 - 186
  • [34] Testing for high-dimensional network parameters in auto-regressive models
    Zheng, Lili
    Raskutti, Garvesh
    ELECTRONIC JOURNAL OF STATISTICS, 2019, 13 (02): : 4977 - 5043
  • [35] Auto-regressive extractive summarization with replacement
    Tianyu Zhu
    Wen Hua
    Jianfeng Qu
    Saeid Hosseini
    Xiaofang Zhou
    World Wide Web, 2023, 26 : 2003 - 2026
  • [36] AN ALGORITHM FOR THE ESTIMATION OF PARAMETERS OF ARMA (AUTO-REGRESSIVE MOVING AVERAGE) MODELS
    DONCARLI, C
    RAIRO-AUTOMATIQUE-SYSTEMS ANALYSIS AND CONTROL, 1982, 16 (01): : 39 - 48
  • [37] Bootstrapping the portmanteau tests in weak auto-regressive moving average models
    Zhu, Ke
    JOURNAL OF THE ROYAL STATISTICAL SOCIETY SERIES B-STATISTICAL METHODOLOGY, 2016, 78 (02) : 463 - 485
  • [38] Spatial auto-correlation and auto-regressive models estimation from sample survey data
    Benedetti, Roberto
    Suesse, Thomas
    Piersimoni, Federica
    BIOMETRICAL JOURNAL, 2020, 62 (06) : 1494 - 1507
  • [39] Dual multivariate auto-regressive modeling in state space for temporal signal separation
    Cheung, YM
    Xu, L
    IEEE TRANSACTIONS ON SYSTEMS MAN AND CYBERNETICS PART B-CYBERNETICS, 2003, 33 (03): : 386 - 398
  • [40] Disentangled Self-Attention with Auto-Regressive Contrastive Learning for Neural Group Recommendation
    Gao, Linyao
    Zhang, Haonan
    Fu, Luoyi
    APPLIED SCIENCES-BASEL, 2024, 14 (10):