A context-enhanced sentence representation learning method for close domains with topic modeling

被引:2
|
作者
Li, Shuangyin [1 ]
Chen, Weiwei [2 ]
Zhang, Yu [3 ]
Zhao, Gansen [1 ]
Pan, Rong [2 ]
Huang, Zhenhua [1 ]
Tang, Yong [1 ]
机构
[1] South China Normal Univ, Sch Comp Sci, Guangzhou, Guangdong, Peoples R China
[2] Sun Yat sen Univ, Sch Data & Comp Sci, Guangzhou, Guangdong, Peoples R China
[3] Southern Univ Sci & Technol, Dept Comp Sci & Engn, Shenzhen, Guangdong, Peoples R China
基金
中国国家自然科学基金;
关键词
Sentence representations learning; Closed domains; Bayesian sentence embedding; Bi-directional context-enhanced; Semantic interpretability; Topic modeling; SHORT TEXT;
D O I
10.1016/j.ins.2022.05.113
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
Sentence representation approaches have been widely used and proven to be effective in many text modeling tasks and downstream applications. Many recent proposals are avail-able on learning sentence representations based on deep neural frameworks. However, these methods are pre-trained in open domains and depend on the availability of large-scale data for model fitting. As a result, they may fail in some special scenarios, where data are sparse and embedding interpretations are required, such as legal, medical, or technical fields. In this paper, we present an unsupervised learning method to exploit representa-tions of sentences for some closed domains via topic modeling. We reformulate the infer-ence process of the sentences with the corresponding contextual sentences and the associated words, and propose an effective context-enhanced process called the bi-Directional Context-enhanced Sentence Representation Learning (bi-DCSR). This method takes advantage of the semantic distributions of the nearby contextual sentences and the associated words to form a context-enhanced sentence representation. To support the bi-DCSR, we develop a novel Bayesian topic model to embed sentences and words into the same latent interpretable topic space called the Hybrid Priors Topic Model (HPTM). Based on the defined topic space by the HPTM, the bi-DCSR method learns the embedding of a sentence by the two-directional contextual sentences and the words in it, which allows us to efficiently learn high-quality sentence representations in such closed domains. In addition to an open-domain dataset from Wikipedia, our method is validated using three closed-domain datasets from legal cases, electronic medical records, and technical reports. Our experiments indicate that the HPTM significantly outperforms on language modeling and topic coherence, compared with the existing topic models. Meanwhile, the bi-DCSR method does not only outperform the state-of-the-art unsupervised learning methods on closed domain sentence classification tasks, but also yields competitive performance com-pared to these established approaches on the open domain. Additionally, the visualizations of the semantics of sentences and words demonstrate the interpretable capacity of our model.(c) 2022 The Author(s). Published by Elsevier Inc. This is an open access article under the CC BY-NC-ND license (http://creativecommons.org/licenses/by-nc-nd/4.0/).
引用
收藏
页码:186 / 210
页数:25
相关论文
共 50 条
  • [1] Context-Enhanced Representation Learning for Single Image Deraining
    Wang, Guoqing
    Sun, Changming
    Sowmya, Arcot
    INTERNATIONAL JOURNAL OF COMPUTER VISION, 2021, 129 (05) : 1650 - 1674
  • [2] Context-Enhanced Representation Learning for Single Image Deraining
    Guoqing Wang
    Changming Sun
    Arcot Sowmya
    International Journal of Computer Vision, 2021, 129 : 1650 - 1674
  • [3] Enhanced Topic Modeling with Multi-modal Representation Learning
    Zhang, Duoyi
    Wang, Yue
    Abul Bashar, Md
    Nayak, Richi
    ADVANCES IN KNOWLEDGE DISCOVERY AND DATA MINING, PAKDD 2023, PT I, 2023, 13935 : 393 - 404
  • [4] Extracting problem and method sentence from scientific papers: a context-enhanced transformer using formulaic expression desensitization
    Zhang, Yingyi
    Zhang, Chengzhi
    SCIENTOMETRICS, 2024, 129 (06) : 3433 - 3468
  • [5] Context-enhanced motion coherence modeling for global outlier rejection
    Li, Hongjie
    Dong, Mingyue
    Zheng, Xianwei
    Xu, Xiong
    Xie, Xiao
    Xiong, Hanjiang
    ISPRS JOURNAL OF PHOTOGRAMMETRY AND REMOTE SENSING, 2023, 202 : 69 - 86
  • [6] Context-Enhanced Vehicle Tracking Method Under the Connected Environment
    Tian, Zhen
    Li, Yinguo
    Cen, Ming
    Zhu, Hao
    Kirubarajan, Thia
    2019 22ND INTERNATIONAL CONFERENCE ON INFORMATION FUSION (FUSION 2019), 2019,
  • [7] Two-Layer Context-Enhanced Representation for Better Chinese Discourse Parsing
    Zhu, Qiang
    Wang, Kedong
    Kong, Fang
    NATURAL LANGUAGE PROCESSING AND CHINESE COMPUTING, NLPCC 2022, PT I, 2022, 13551 : 43 - 54
  • [8] Regularized and Retrofitted models for Learning Sentence Representation with Context
    Saha, Tanay Kumar
    Joty, Shafiq
    Hassan, Naeemul
    Al Hasan, Mohammad
    CIKM'17: PROCEEDINGS OF THE 2017 ACM CONFERENCE ON INFORMATION AND KNOWLEDGE MANAGEMENT, 2017, : 547 - 556
  • [9] Ranking-Enhanced Unsupervised Sentence Representation Learning
    Seonwoo, Yeon
    Wang, Guoyin
    Seo, Changmin
    Choudhary, Sajal
    Li, Jiwei
    Li, Xiang
    Xu, Puyang
    Park, Sunghyun
    Oh, Alice
    PROCEEDINGS OF THE 61ST ANNUAL MEETING OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS (ACL 2023): LONG PAPERS, VOL 1, 2023, : 15783 - 15798
  • [10] CETA: Context-Enhanced and Target-Aware Hateful Meme Inference Method
    Wang, Kaichun
    Lu, Junyu
    Yu, Bingjie
    Yang, Liang
    Lin, Hongfei
    NATURAL LANGUAGE PROCESSING AND CHINESE COMPUTING, PT V, NLPCC 2024, 2025, 15363 : 95 - 106