A context-enhanced sentence representation learning method for close domains with topic modeling

被引:2
|
作者
Li, Shuangyin [1 ]
Chen, Weiwei [2 ]
Zhang, Yu [3 ]
Zhao, Gansen [1 ]
Pan, Rong [2 ]
Huang, Zhenhua [1 ]
Tang, Yong [1 ]
机构
[1] South China Normal Univ, Sch Comp Sci, Guangzhou, Guangdong, Peoples R China
[2] Sun Yat sen Univ, Sch Data & Comp Sci, Guangzhou, Guangdong, Peoples R China
[3] Southern Univ Sci & Technol, Dept Comp Sci & Engn, Shenzhen, Guangdong, Peoples R China
基金
中国国家自然科学基金;
关键词
Sentence representations learning; Closed domains; Bayesian sentence embedding; Bi-directional context-enhanced; Semantic interpretability; Topic modeling; SHORT TEXT;
D O I
10.1016/j.ins.2022.05.113
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
Sentence representation approaches have been widely used and proven to be effective in many text modeling tasks and downstream applications. Many recent proposals are avail-able on learning sentence representations based on deep neural frameworks. However, these methods are pre-trained in open domains and depend on the availability of large-scale data for model fitting. As a result, they may fail in some special scenarios, where data are sparse and embedding interpretations are required, such as legal, medical, or technical fields. In this paper, we present an unsupervised learning method to exploit representa-tions of sentences for some closed domains via topic modeling. We reformulate the infer-ence process of the sentences with the corresponding contextual sentences and the associated words, and propose an effective context-enhanced process called the bi-Directional Context-enhanced Sentence Representation Learning (bi-DCSR). This method takes advantage of the semantic distributions of the nearby contextual sentences and the associated words to form a context-enhanced sentence representation. To support the bi-DCSR, we develop a novel Bayesian topic model to embed sentences and words into the same latent interpretable topic space called the Hybrid Priors Topic Model (HPTM). Based on the defined topic space by the HPTM, the bi-DCSR method learns the embedding of a sentence by the two-directional contextual sentences and the words in it, which allows us to efficiently learn high-quality sentence representations in such closed domains. In addition to an open-domain dataset from Wikipedia, our method is validated using three closed-domain datasets from legal cases, electronic medical records, and technical reports. Our experiments indicate that the HPTM significantly outperforms on language modeling and topic coherence, compared with the existing topic models. Meanwhile, the bi-DCSR method does not only outperform the state-of-the-art unsupervised learning methods on closed domain sentence classification tasks, but also yields competitive performance com-pared to these established approaches on the open domain. Additionally, the visualizations of the semantics of sentences and words demonstrate the interpretable capacity of our model.(c) 2022 The Author(s). Published by Elsevier Inc. This is an open access article under the CC BY-NC-ND license (http://creativecommons.org/licenses/by-nc-nd/4.0/).
引用
收藏
页码:186 / 210
页数:25
相关论文
共 50 条
  • [31] Enhanced context-aware recommendation using topic modeling and particle swarm optimization
    Gasmi, Ibtissem
    Azizi, Mohamed Walid
    Seridi-Bouchelaghem, Hassina
    Azizi, Nabiha
    Belhaouari, Samir Brahim
    JOURNAL OF INTELLIGENT & FUZZY SYSTEMS, 2021, 40 (06) : 12227 - 12242
  • [32] Representation learning approach for medical activities enhanced by topical modeling
    Xu X.
    Wang Y.
    Jin T.
    Wang J.
    Qinghua Daxue Xuebao/Journal of Tsinghua University, 2019, 59 (03): : 169 - 177
  • [33] Speeding up Context-based Sentence Representation Learning with Non-autoregressive Convolutional Decoding
    Tang, Shuai
    Jin, Hailin
    Fang, Chen
    Wang, Zhaowen
    de Sa, Virginia R.
    REPRESENTATION LEARNING FOR NLP, 2018, : 69 - 78
  • [34] Development of a Method for Enhanced Fan Representation in Gas Turbine Modeling
    Doulgeris, Georgios
    Khaleghi, Hossein
    Kalfas, Anestis
    Pilidis, Pericles
    INTERNATIONAL JOURNAL OF ROTATING MACHINERY, 2011, 2011
  • [35] Joint Representation Learning with Relation-Enhanced Topic Models for Intelligent Job Interview Assessment
    Shen, Dazhong
    Qin, Chuan
    Zhu, Hengshu
    Xu, Tong
    Chen, Enhong
    Xiong, Hui
    ACM TRANSACTIONS ON INFORMATION SYSTEMS, 2022, 40 (01)
  • [36] Semi-supervised Topic Learning and Representation Method Based on Association Rules and Metadata
    Zhao Huiru
    Lin Min
    2017 2ND IEEE INTERNATIONAL CONFERENCE ON COMPUTATIONAL INTELLIGENCE AND APPLICATIONS (ICCIA), 2017, : 51 - 56
  • [37] Adaptive and hybrid context-aware fine-grained word sense disambiguation in topic modeling based document representation
    Li, Wenbo
    Suzuki, Einoshin
    Information Processing and Management, 2021, 58 (04):
  • [38] Adaptive and hybrid context-aware fine-grained word sense disambiguation in topic modeling based document representation
    Li, Wenbo
    Suzuki, Einoshin
    INFORMATION PROCESSING & MANAGEMENT, 2021, 58 (04)
  • [39] Unsupervised speech representation learning for behavior modeling using triplet enhanced contextualized networks
    Li, Haoqi
    Baucom, Brian
    Narayanan, Shrikanth
    Georgiou, Panayiotis
    COMPUTER SPEECH AND LANGUAGE, 2021, 70
  • [40] RLRecommender: A Representation-Learning-Based Recommendation Method for Business Process Modeling
    Wang, Huaqing
    Wen, Lijie
    Lin, Li
    Wang, Jianmin
    SERVICE-ORIENTED COMPUTING (ICSOC 2018), 2018, 11236 : 478 - 486