On Partial Multi-Task Learning

被引:5
|
作者
He, Yi [1 ]
Wu, Baijun [1 ]
Wu, Di [2 ]
Wu, Xindong [3 ,4 ]
机构
[1] Univ Louisiana Lafayette, Sch Comp & Informat, Lafayette, LA 70504 USA
[2] Chinese Acad Sci, Chongqing Inst Green & Intelligent Technol, Beijing, Peoples R China
[3] Mininglamp Acad Sci, Mininglamp Technol, Beijing, Peoples R China
[4] Hefei Univ Technol, Minist Educ, Key Lab Knowledge Engn Big Data, Hefei, Peoples R China
基金
美国国家科学基金会;
关键词
MATRIX COMPLETION; CLASSIFICATION;
D O I
10.3233/FAIA200216
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Multi-Task Learning (MTL) has shown its effectiveness in real applications where many related tasks could be handled together. Existing MTL methods make predictions for multiple tasks based on the data examples of the corresponding tasks. However, the data examples of some tasks are expensive or time-consuming to collect in practice, which reduces the applicability of MTL. For example, a patient may be asked to provide her microtome test reports and MRI images for illness diagnosis in MTL-based system [37,40]. It would be valuable if MTL can predict the abnormalities for such medical tests by feeding with some easy-to-collect data examples from other related tests instead of directly collecting data examples from them. We term such a new paradigm as multi-task learning from partial examples. The challenges of partial multi-task learning are twofold. First, the data examples from different tasks may be represented in different feature spaces. Second, the data examples could be incomplete for predicting the labels of all tasks. To overcome these challenges, we in this paper propose a novel algorithm, named Generative Learning with Partial Multi-Tasks (GPMT). The key idea of GPMT is to discover a shared latent feature space that harmonizes disparate feature information of multiple tasks. Given a partial example, the information contained in its missing feature representations is recovered by projecting it onto the latent space. A learner trained on the latent space then enjoys complete information included in the original features and the recovered missing features, and thus can predict the labels for the partial examples. Our theoretical analysis shows that the GPMT guarantees a performance gain comparing with training an individual learner for each task. Extensive experiments demonstrate the superiority of GPMT on both synthetic and real datasets.
引用
收藏
页码:1174 / 1181
页数:8
相关论文
共 50 条
  • [41] Polymer informatics with multi-task learning
    Kuenneth, Christopher
    Rajan, Arunkumar Chitteth
    Tran, Huan
    Chen, Lihua
    Kim, Chiho
    Ramprasad, Rampi
    PATTERNS, 2021, 2 (04):
  • [42] Kernel Online Multi-task Learning
    Sumitra, S.
    Aravindh, A.
    COMPUTATIONAL INTELLIGENCE, CYBER SECURITY AND COMPUTATIONAL MODELS, ICC3 2015, 2016, 412 : 55 - 64
  • [43] Multi-task Learning for Recommender Systems
    Ning, Xia
    Karypis, George
    PROCEEDINGS OF 2ND ASIAN CONFERENCE ON MACHINE LEARNING (ACML2010), 2010, 13 : 269 - 284
  • [44] Stock Ranking with Multi-Task Learning
    Ma, Tao
    Tan, Ying
    EXPERT SYSTEMS WITH APPLICATIONS, 2022, 199
  • [45] Bounds for linear multi-task learning
    Maurer, A
    JOURNAL OF MACHINE LEARNING RESEARCH, 2006, 7 : 117 - 139
  • [46] ADAPTIVE AND ROBUST MULTI-TASK LEARNING
    Duan, Yaqi
    Wang, Kaizheng
    ANNALS OF STATISTICS, 2023, 51 (05): : 2015 - 2039
  • [47] Multi-task learning for pKa prediction
    Skolidis, Grigorios
    Hansen, Katja
    Sanguinetti, Guido
    Rupp, Matthias
    JOURNAL OF COMPUTER-AIDED MOLECULAR DESIGN, 2012, 26 (07) : 883 - 895
  • [48] Federated Multi-task Graph Learning
    Liu, Yijing
    Han, Dongming
    Zhang, Jianwei
    Zhu, Haiyang
    Xu, Mingliang
    Chen, Wei
    ACM TRANSACTIONS ON INTELLIGENT SYSTEMS AND TECHNOLOGY, 2022, 13 (05)
  • [49] Multi-task and Lifelong Learning of Kernels
    Pentina, Anastasia
    Ben-David, Shai
    ALGORITHMIC LEARNING THEORY, ALT 2015, 2015, 9355 : 194 - 208
  • [50] Deep Decentralized Multi-task Multi-Agent Reinforcement Learning under Partial Observability
    Omidshafiei, Shayegan
    Pazis, Jason
    Amato, Christopher
    How, Jonathan P.
    Vian, John
    INTERNATIONAL CONFERENCE ON MACHINE LEARNING, VOL 70, 2017, 70