MTLAN: Multi-Task Learning and Auxiliary Network for Enhanced Sentence Embedding

被引:0
|
作者
Liu, Gang [1 ,2 ]
Wang, Tongli [1 ]
Yang, Wenli [1 ]
Yan, Zhizheng [1 ]
Zhan, Kai [3 ]
机构
[1] Harbin Engn Univ, Coll Comp Sci & Technol, Harbin, Peoples R China
[2] Harbin Engn Univ, Modeling & Emulat E Govt Natl Engn Lab, Harbin, Peoples R China
[3] PwC Enterprise Digital, PricewaterhouseCoopers, Sydney, NSW, Australia
关键词
Cross-lingual; Sentence embedding; Multi-task learning; Contrastive learning; Auxiliary network;
D O I
10.1007/978-981-99-8067-3_2
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
The objective of cross-lingual sentence embedding learning is to map sentences into a shared representation space, where semantically similar sentence representations are closer together, while distinct sentence representations exhibit clear differentiation. This paper proposes a novel sentence embedding model called MTLAN, which incorporates multi-task learning and auxiliary networks. The model utilizes the LaBSE model for extracting sentence features and undergoes joint training on tasks related to sentence semantic representation and distance measurement. Furthermore, an auxiliary network is employed to enhance the contextual expression of words within sentences. To address the issue of limited resources for low-resource languages, we construct a pseudocorpus dataset using a multilingual dictionary for unsupervised learning. We conduct experiments on multiple publicly available datasets, including STS and SICK, to evaluate both monolingual sentence similarity and cross-lingual semantic similarity. The empirical results demonstrate the significant superiority of our proposed model over state-of-the-art methods.
引用
收藏
页码:16 / 27
页数:12
相关论文
共 50 条
  • [1] Multi-Task Learning Based Network Embedding
    Wang, Shanfeng
    Wang, Qixiang
    Gong, Maoguo
    FRONTIERS IN NEUROSCIENCE, 2020, 13
  • [2] Cross-lingual Sentence Embedding using Multi-Task Learning
    Goswami, Koustava
    Dutta, Sourav
    Assem, Haytham
    Fransen, Theodorus
    McCrae, John P.
    2021 CONFERENCE ON EMPIRICAL METHODS IN NATURAL LANGUAGE PROCESSING (EMNLP 2021), 2021, : 9099 - 9113
  • [3] SEBGM: Sentence Embedding Based on Generation Model with multi-task learning
    Wang, Qian
    Zhang, Weiqi
    Lei, Tianyi
    Cao, Yu
    Peng, Dezhong
    Wang, Xu
    COMPUTER SPEECH AND LANGUAGE, 2024, 87
  • [4] Multi-task Network Embedding
    Xu, Linchuan
    Wei, Xiaokai
    Cao, Jiannong
    Yu, Philip S.
    2017 IEEE INTERNATIONAL CONFERENCE ON DATA SCIENCE AND ADVANCED ANALYTICS (DSAA), 2017, : 571 - 580
  • [5] Multi-task network embedding
    Linchuan Xu
    Xiaokai Wei
    Jiannong Cao
    Philip S. Yu
    International Journal of Data Science and Analytics, 2019, 8 : 183 - 198
  • [6] Multi-task network embedding
    Xu, Linchuan
    Wei, Xiaokai
    Cao, Jiannong
    Yu, Philip S.
    INTERNATIONAL JOURNAL OF DATA SCIENCE AND ANALYTICS, 2019, 8 (02) : 183 - 198
  • [7] Task Switching Network for Multi-task Learning
    Sun, Guolei
    Probst, Thomas
    Paudel, Danda Pani
    Popovic, Nikola
    Kanakis, Menelaos
    Patel, Jagruti
    Dai, Dengxin
    Van Gool, Luc
    2021 IEEE/CVF INTERNATIONAL CONFERENCE ON COMPUTER VISION (ICCV 2021), 2021, : 8271 - 8280
  • [8] Multi-task Network Embedding with Adaptive Loss Weighting
    Rizi, Fatemeh Salehi
    Granitzer, Michael
    2020 IEEE/ACM INTERNATIONAL CONFERENCE ON ADVANCES IN SOCIAL NETWORKS ANALYSIS AND MINING (ASONAM), 2020, : 1 - 5
  • [9] Hierarchical multi-task learning withself-supervised auxiliary task
    Lee, Seunghan
    Park, Taeyoung
    KOREAN JOURNAL OF APPLIED STATISTICS, 2024, 37 (05)
  • [10] Unified Voice Embedding through Multi-task Learning
    Rajenthiran, Jenarthanan
    Sithamaparanathan, Lakshikka
    Uthayakumar, Saranya
    Thayasivam, Uthayasanker
    2022 INTERNATIONAL CONFERENCE ON ASIAN LANGUAGE PROCESSING (IALP 2022), 2022, : 178 - 183