Accelerated Gradient Method for Multi-Task Sparse Learning Problem

被引:126
|
作者
Chen, Xi [1 ]
Pan, Weike [2 ]
Kwok, James T. [2 ]
Carbonell, Jaime G. [1 ]
机构
[1] Carnegie Mellon Univ, Sch Comp Sci, Pittsburgh, PA 15213 USA
[2] Hong Kong Univ Sci & Technol, Dept Comp Sci & Technol, Hong Kong, Peoples R China
关键词
multi-task learning; L-1-infinity regularization; optimal method; gradient descend; SHRINKAGE;
D O I
10.1109/ICDM.2009.128
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
Many real world learning problems can be recast as multi-task learning problems which utilize correlations among different tasks to obtain better generalization performance than learning each task individually. The feature selection problem in multi-task setting has many applications in fields of computer vision, text classification and bio-informatics. Generally, it can be realized by solving a L-1-infinity regularized optimization problem. And the solution automatically yields the joint sparsity among different tasks. However, due to the nonsmooth nature of the L-1-infinity norm, there lacks an efficient training algorithm for solving such problem with general convex loss functions. In this paper, we propose an accelerated gradient method based on an "optimal" first order black-box method named after Nesterov and provide the convergence rate for smooth convex loss functions. For nonsmooth convex loss functions, such as hinge loss, our method still has fast convergence rate empirically. Moreover, by exploiting the structure of the L-1-infinity ball, we solve the black-box oracle in Nesterov's method by a simple sorting scheme. Our method is suitable for large-scale multi-task learning problem since it only utilizes the first order information and is very easy to implement. Experimental results show that our method significantly outperforms the most state-of-the-art methods in both convergence speed and learning accuracy.
引用
收藏
页码:746 / +
页数:2
相关论文
共 50 条
  • [1] Multi-task gradient descent for multi-task learning
    Lu Bai
    Yew-Soon Ong
    Tiantian He
    Abhishek Gupta
    Memetic Computing, 2020, 12 : 355 - 369
  • [2] Multi-task gradient descent for multi-task learning
    Bai, Lu
    Ong, Yew-Soon
    He, Tiantian
    Gupta, Abhishek
    MEMETIC COMPUTING, 2020, 12 (04) : 355 - 369
  • [3] Learning Sparse Task Relations in Multi-Task Learning
    Zhang, Yu
    Yang, Qiang
    THIRTY-FIRST AAAI CONFERENCE ON ARTIFICIAL INTELLIGENCE, 2017, : 2914 - 2920
  • [4] Sparse Multi-Task Reinforcement Learning
    Calandriello, Daniele
    Lazaric, Alessandro
    Restelli, Marcello
    ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 27 (NIPS 2014), 2014, 27
  • [5] Sparse multi-task reinforcement learning
    Calandriello, Daniele
    Lazaric, Alessandro
    Restelli, Marcello
    INTELLIGENZA ARTIFICIALE, 2015, 9 (01) : 5 - 20
  • [6] Gradient Surgery for Multi-Task Learning
    Yu, Tianhe
    Kumar, Saurabh
    Gupta, Abhishek
    Levine, Sergey
    Hausman, Karol
    Finn, Chelsea
    ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 33, NEURIPS 2020, 2020, 33
  • [7] Online Multi-Task Learning for Policy Gradient Methods
    Ammar, Haitham Bou
    Eaton, Eric
    Ruvolo, Paul
    Taylor, Matthew E.
    INTERNATIONAL CONFERENCE ON MACHINE LEARNING, VOL 32 (CYCLE 2), 2014, 32 : 1206 - 1214
  • [8] Compressive sensing based visual tracking using multi-task sparse learning method
    Kang, Bin
    Zhang, Ling-Hua
    Zhu, Wei-Ping
    Lun, Daniel Pak Kong
    2016 8TH INTERNATIONAL CONFERENCE ON WIRELESS COMMUNICATIONS & SIGNAL PROCESSING (WCSP), 2016,
  • [9] Tackling ordinal regression problem for heterogeneous data: sparse and deep multi-task learning approaches
    Lu Wang
    Dongxiao Zhu
    Data Mining and Knowledge Discovery, 2021, 35 : 1134 - 1161
  • [10] Tackling ordinal regression problem for heterogeneous data: sparse and deep multi-task learning approaches
    Wang, Lu
    Zhu, Dongxiao
    DATA MINING AND KNOWLEDGE DISCOVERY, 2021, 35 (03) : 1134 - 1161