Multi-Task Learning in Natural Language Processing: An Overview

被引:10
|
作者
Chen, Shijie [1 ]
Zhang, Yu [2 ]
Yang, Qiang [3 ]
机构
[1] Ohio State Univ, Dept Comp Sci & Engn, Columbus, OH USA
[2] Southern Univ Sci & Technol, Dept Comp Sci & Engn, Shenzhen, Peoples R China
[3] Hong Kong Univ Sci & Technol, Dept Comp Sci & Engn, Hong Kong, Peoples R China
关键词
Multi-task learning;
D O I
10.1145/3663363
中图分类号
TP301 [理论、方法];
学科分类号
081202 ;
摘要
Deep learning approaches have achieved great success in the field of Natural Language Processing (NLP). However, directly training deep neural models often suffer from overfitting and data scarcity problems that are pervasive in NLP tasks. In recent years, Multi-Task Learning (MTL), which can leverage useful information of related tasks to achieve simultaneous performance improvement on these tasks, has been used to handle these problems. In this article, we give an overview of the use of MTL in NLP tasks. We first review MTL architectures used in NLP tasks and categorize them into four classes, including parallel architecture, hierarchical architecture, modular architecture, and generative adversarial architecture. Then we present optimization techniques on loss construction, gradient regularization, data sampling, and task scheduling to properly train a multi-task model. After presenting applications of MTL in a variety of NLP tasks, we introduce some benchmark datasets. Finally, we make a conclusion and discuss several possible research directions in this field.
引用
收藏
页数:32
相关论文
共 50 条
  • [1] Survey of Multi-task Learning in Natural Language Processing: Regarding Task Relatedness and Training Methods
    Zhang, Zhihan
    Yu, Wenhao
    Yu, Mengxia
    Guo, Zhichun
    Jiang, Meng
    17TH CONFERENCE OF THE EUROPEAN CHAPTER OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS, EACL 2023, 2023, : 943 - 956
  • [2] An overview of multi-task learning
    Zhang, Yu
    Yang, Qiang
    NATIONAL SCIENCE REVIEW, 2018, 5 (01) : 30 - 43
  • [3] An overview of multi-task learning
    Yu Zhang
    Qiang Yang
    National Science Review, 2018, 5 (01) : 30 - 43
  • [4] Empirical evaluation of multi-task learning in deep neural networks for natural language processing
    Jianquan Li
    Xiaokang Liu
    Wenpeng Yin
    Min Yang
    Liqun Ma
    Yaohong Jin
    Neural Computing and Applications, 2021, 33 : 4417 - 4428
  • [5] Empirical evaluation of multi-task learning in deep neural networks for natural language processing
    Li, Jianquan
    Liu, Xiaokang
    Yin, Wenpeng
    Yang, Min
    Ma, Liqun
    Jin, Yaohong
    NEURAL COMPUTING & APPLICATIONS, 2021, 33 (09): : 4417 - 4428
  • [6] A Multi-Task Semantic Communication System for Natural Language Processing
    Sheng, Yucheng
    Li, Fang
    Liang, Le
    Jin, Shi
    2022 IEEE 96TH VEHICULAR TECHNOLOGY CONFERENCE (VTC2022-FALL), 2022,
  • [7] Multi-task learning for natural language processing in the 2020s: Where are we going?
    Worsham, Joseph
    Kalita, Jugal
    PATTERN RECOGNITION LETTERS, 2020, 136 (136) : 120 - 126
  • [8] Multi-task Learning for Natural Language Generation in Task-Oriented Dialogue
    Zhu, Chenguang
    Zeng, Michael
    Huang, Xuedong
    2019 CONFERENCE ON EMPIRICAL METHODS IN NATURAL LANGUAGE PROCESSING AND THE 9TH INTERNATIONAL JOINT CONFERENCE ON NATURAL LANGUAGE PROCESSING (EMNLP-IJCNLP 2019): PROCEEDINGS OF THE CONFERENCE, 2019, : 1261 - 1266
  • [9] Bidirectional Transformer Based Multi-Task Learning for Natural Language Understanding
    Tripathi, Suraj
    Singh, Chirag
    Kumar, Abhay
    Pandey, Chandan
    Jain, Nishant
    NATURAL LANGUAGE PROCESSING AND INFORMATION SYSTEMS (NLDB 2019), 2019, 11608 : 54 - 65
  • [10] TaskFusion: An Efficient Transfer Learning Architecture with Dual Delta Sparsity for Multi-Task Natural Language Processing
    Fan, Zichen
    Zhang, Qirui
    Abillama, Pierre
    Shoouri, Sara
    Lee, Changwoo
    Blaauw, David
    Kim, Hun-Seok
    Sylvester, Dennis
    PROCEEDINGS OF THE 2023 THE 50TH ANNUAL INTERNATIONAL SYMPOSIUM ON COMPUTER ARCHITECTURE, ISCA 2023, 2023, : 62 - 75