TC-DWA: Text Clustering with Dual Word-Level Augmentation

被引:0
|
作者
Cheng, Bo [1 ,4 ,5 ]
Li, Ximing [2 ,3 ]
Chang, Yi [1 ,4 ,5 ]
机构
[1] Jilin Univ, Sch Artificial Intelligence, Jilin, Jilin, Peoples R China
[2] Jilin Univ, Coll Comp Sci & Technol, Jilin, Jilin, Peoples R China
[3] Jilin Univ, Key Lab Symbol Computat & Knowledge Engn, MOE, Jilin, Jilin, Peoples R China
[4] Jilin Univ, Int Ctr Future Sci, Jilin, Jilin, Peoples R China
[5] Minist Educ, Engn Res Ctr Knowledge Driven Human Machine Intel, Beijing, Peoples R China
基金
中国国家自然科学基金; 国家重点研发计划;
关键词
D O I
暂无
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
The pre-trained language models, e.g., ELMo and BERT, have recently achieved promising performance improvement in a wide range of NLP tasks, because they can output strong contextualized embedded features of words. Inspired by their great success, in this paper we aim to fine-tune them to effectively handle the text clustering task, i.e., a classic and fundamental challenge in machine learning. Accordingly, we propose a novel BERT-based method, namely Text Clustering with Dual Word-level Augmentation (TC-DWA). To be specific, we formulate a self-training objective and enhance it with a dual word-level augmentation technique. First, we suppose that each text contains several most informative words, called anchor words, supporting the full text semantics. We use the embedded features of anchor words as augmented features, which are selected by ranking the norm-based attention weights of words. Second, we formulate an expectation form of word augmentation, which is equivalent to generating infinite augmented features, and further suggest a tractable approximation of Taylor expansion for efficient optimization. To evaluate the effectiveness of TC-DWA, we conduct extensive experiments on several benchmark text datasets. The results demonstrate that TC-DWA consistently outperforms the state-of-the-art baseline methods. Code available: https://github.com/BoCheng-96/TC-DWA.
引用
收藏
页码:7113 / 7121
页数:9
相关论文
共 37 条
  • [21] Relation Extraction in Vietnamese Text via Piecewise Convolution Neural Network with Word-Level Attention
    Van-Nhat Nguyen
    Ha-Thanh Nguyen
    Dinh-Hieu Vo
    Le-Minh Nguyen
    PROCEEDINGS OF 2018 5TH NAFOSTED CONFERENCE ON INFORMATION AND COMPUTER SCIENCE (NICS 2018), 2018, : 99 - 103
  • [22] Word-level human interpretable scoring mechanism for novel text detection using Tsetlin Machines
    Bimal Bhattarai
    Ole-Christoffer Granmo
    Lei Jiao
    Applied Intelligence, 2022, 52 : 17465 - 17489
  • [23] Word-level human interpretable scoring mechanism for novel text detection using Tsetlin Machines
    Bhattarai, Bimal
    Granmo, Ole-Christoffer
    Jiao, Lei
    APPLIED INTELLIGENCE, 2022, 52 (15) : 17465 - 17489
  • [24] Relative effectiveness of reading practice or word-level instruction in supplemental tutoring: How text matters
    Vadasy, PF
    Sanders, EA
    Peyton, JA
    JOURNAL OF LEARNING DISABILITIES, 2005, 38 (04) : 364 - 380
  • [25] Cross-language message- and word-level transfer effects in bilingual text processing
    Deanna C. Friesen
    Debra Jared
    Memory & Cognition, 2007, 35 : 1542 - 1556
  • [26] Cross-language message- and word-level transfer effects in bilingual text processing
    Friesen, Deanna C.
    Jared, Debra
    MEMORY & COGNITION, 2007, 35 (07) : 1542 - 1556
  • [27] Contrasting Human- and Machine-Generated Word-Level Adversarial Examples for Text Classification
    Mozes, Maximilian
    Bartolo, Max
    Stenetorp, Pontus
    Kleinberg, Bennett
    Griffin, Lewis D.
    2021 CONFERENCE ON EMPIRICAL METHODS IN NATURAL LANGUAGE PROCESSING (EMNLP 2021), 2021, : 8258 - 8270
  • [28] Semi-Supervised Multinomial Naive Bayes for Text Classification by Leveraging Word-Level Statistical Constraint
    Zhao, Li
    Huang, Minlie
    Yao, Ziyu
    Su, Rongwei
    Jiang, Yingying
    Zhu, Xiaoyan
    THIRTIETH AAAI CONFERENCE ON ARTIFICIAL INTELLIGENCE, 2016, : 2877 - 2883
  • [29] Enhancing Word-Level Semantic Representation via Dependency Structure for Expressive Text-to-Speech Synthesis
    Zhou, Yixuan
    Song, Changhe
    Li, Jingbei
    Wu, Zhiyong
    Bian, Yanyao
    Su, Dan
    Meng, Helen
    INTERSPEECH 2022, 2022, : 5518 - 5522
  • [30] SwitchNet: Learning to switch for word-level language identification in code-mixed social media text
    Sarma, Neelakshi
    Sanasam Singh, Ranbir
    Goswami, Diganta
    NATURAL LANGUAGE ENGINEERING, 2022, 28 (03) : 337 - 359