A new transfer learning framework with application to model-agnostic multi-task learning

被引:7
|
作者
Gupta, Sunil [1 ]
Rana, Santu [1 ]
Saha, Budhaditya [1 ]
Phung, Dinh [1 ]
Venkatesh, Svetha [1 ]
机构
[1] Deakin Univ, Ctr Pattern Recognit & Data Analyt PRaDA, Geelong Waurn Ponds Campus, Waurn Ponds, Vic, Australia
关键词
Multi-task learning; Model-agnostic framework; Meta algorithm; Classification; Regression; CLASSIFICATION;
D O I
10.1007/s10115-016-0926-z
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Learning from small number of examples is a challenging problem in machine learning. An effective way to improve the performance is through exploiting knowledge from other related tasks. Multi-task learning (MTL) is one such useful paradigm that aims to improve the performance through jointly modeling multiple related tasks. Although there exist numerous classification or regression models in machine learning literature, most of the MTL models are built around ridge or logistic regression. There exist some limited works, which propose multi-task extension of techniques such as support vector machine, Gaussian processes. However, all these MTL models are tied to specific classification or regression algorithms and there is no single MTL algorithm that can be used at a meta level for any given learning algorithm. Addressing this problem, we propose a generic, model-agnostic joint modeling framework that can take any classification or regression algorithm of a practitioner's choice (standard or custom-built) and build its MTL variant. The key observation that drives our framework is that due to small number of examples, the estimates of task parameters are usually poor, and we show that this leads to an under-estimation of task relatedness between any two tasks with high probability. We derive an algorithm that brings the tasks closer to their true relatedness by improving the estimates of task parameters. This is achieved by appropriate sharing of data across tasks. We provide the detail theoretical underpinning of the algorithm. Through our experiments with both synthetic and real datasets, we demonstrate that the multi-task variants of several classifiers/regressors (logistic regression, support vector machine, K-nearest neighbor, Random Forest, ridge regression, support vector regression) convincingly outperform their single-task counterparts. We also show that the proposed model performs comparable or better than many state-of-the-art MTL and transfer learning baselines.
引用
收藏
页码:933 / 973
页数:41
相关论文
共 50 条
  • [1] A new transfer learning framework with application to model-agnostic multi-task learning
    Sunil Gupta
    Santu Rana
    Budhaditya Saha
    Dinh Phung
    Svetha Venkatesh
    Knowledge and Information Systems, 2016, 49 : 933 - 973
  • [2] A Model-Agnostic Approach to Mitigate Gradient Interference for Multi-Task Learning
    Chai, Heyan
    Yin, Zhe
    Ding, Ye
    Liu, Li
    Fang, Binxing
    Liao, Qing
    IEEE TRANSACTIONS ON CYBERNETICS, 2023, 53 (12) : 7810 - 7823
  • [3] MASR: A Model-Agnostic Sparse Routing Architecture for Arbitrary Order Feature Sharing in Multi-Task Learning
    Dong, Xin
    Wu, Ruize
    Xiong, Chao
    Li, Hai
    Cheng, Lei
    He, Yong
    Qian, Shiyou
    Cao, Jian
    Mo, Linjian
    PROCEEDINGS OF THE 31ST ACM INTERNATIONAL CONFERENCE ON INFORMATION AND KNOWLEDGE MANAGEMENT, CIKM 2022, 2022, : 3923 - 3927
  • [4] Episodic task agnostic contrastive training for multi-task learning?
    Zhou, Fan
    Chen, Yuyi
    Wen, Jun
    Zeng, Qiuhao
    Shui, Changjian
    Ling, Charles X.
    Yang, Shichun
    Wang, Boyu
    NEURAL NETWORKS, 2023, 162 : 34 - 45
  • [5] Combining Model-Agnostic Meta-Learning and Transfer Learning for Regression
    Satrya, Wahyu Fadli
    Yun, Ji-Hoon
    SENSORS, 2023, 23 (02)
  • [6] Model-Agnostic Federated Learning
    Mittone, Gianluca
    Riviera, Walter
    Colonnelli, Iacopo
    Birke, Robert
    Aldinucci, Marco
    EURO-PAR 2023: PARALLEL PROCESSING, 2023, 14100 : 383 - 396
  • [7] Model-Agnostic Private Learning
    Bassily, Raef
    Thakkar, Om
    Thakurta, Abhradeep
    ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 31 (NIPS 2018), 2018, 31
  • [8] Task-Robust Model-Agnostic Meta-Learning
    Collins, Liam
    Mokhtari, Aryan
    Shakkottai, Sanjay
    ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 33, NEURIPS 2020, 2020, 33
  • [9] Is Bayesian Model-Agnostic Meta Learning Better than Model-Agnostic Meta Learning, Provably?
    Chen, Lisha
    Chen, Tianyi
    INTERNATIONAL CONFERENCE ON ARTIFICIAL INTELLIGENCE AND STATISTICS, VOL 151, 2022, 151
  • [10] A multi-task transfer learning method with dictionary learning
    Zheng, Xin
    Lin, Luyue
    Liu, Bo
    Xiao, Yanshan
    Xiong, Xiaoming
    KNOWLEDGE-BASED SYSTEMS, 2020, 191