Compression-Based Regularization With an Application to Multitask Learning

被引:6
|
作者
Vera, Matias [1 ]
Vega, Leonardo Rey [1 ,2 ]
Piantanida, Pablo [3 ]
机构
[1] Univ Buenos Aires, Fac Ingn, RA-1053 Buenos Aires, DF, Argentina
[2] Univ Buenos Aires, CSC CONICET, RA-1053 Buenos Aires, DF, Argentina
[3] Univ Paris Sud, CNRS, Cent Supelec, F-91400 Orsay, France
关键词
Multitask learning; information bottleneck; regularization; Arimoto-Blahut algorithm; side information; CLASSIFICATION; INFORMATION; FRAMEWORK; CAPACITY;
D O I
10.1109/JSTSP.2018.2846218
中图分类号
TM [电工技术]; TN [电子技术、通信技术];
学科分类号
0808 ; 0809 ;
摘要
This paper investigates, from information theoretic grounds, a learning problem based on the principle that any regularity in a given dataset can be exploited to extract compact features from data, i.e., using fewer bits than needed to fully describe the data itself, in order to build meaningful representations of a relevant content (multiple labels). We begin studying a multitask learning (MTL) problem from the average (over the tasks) of mis-classification probability point of view and linking it with the popular cross-entropy criterion. Our approach allows an information theoretic formulation of an MTL problem as a supervised learning framework, in which the prediction models for several related tasks are learned jointly from common representations to achieve better generalization performance. More precisely, our formulation of the MTL problem can be interpreted as an information bottleneck problem with side information at the decoder. Based on that, we present an iterative algorithm for computing the optimal tradeoffs and some of its convergence properties are studied. An important feature of this algorithm is to provide a natural safeguard against overfitting, because it minimizes the average risk taking into account a penalization induced by the model complexity. Remarkably, empirical results illustrate that there exists an optimal information rate minimizing the excess risk, which depends on the nature and the amount of available training data. Applications to hierarchical text categorization and distributional word clusters are also investigated, extending previous works.
引用
收藏
页码:1063 / 1076
页数:14
相关论文
共 50 条
  • [21] On the application of compression-based metrics to identifying anomalous behaviour in web traffic
    De La Torre-Abaitua, Gonzalo
    Lago-Fernandez, Luis F.
    Arroyo, David
    LOGIC JOURNAL OF THE IGPL, 2020, 28 (04) : 546 - 557
  • [22] A compression-based distance measure for texture
    Campana B.J.L.
    Keogh E.J.
    Statistical Analysis and Data Mining, 2010, 3 (06): : 381 - 398
  • [23] Regularization-Based Multitask Learning With Applications to Genome Biology and Biological Imaging
    Widmer, Christian
    Kloft, Marius
    Lou, Xinghua
    Raetsch, Gunnar
    KUNSTLICHE INTELLIGENZ, 2014, 28 (01): : 29 - 33
  • [24] Compression-based analysis of metamorphic malware
    Department of Computer Science, San Jose State University, San Jose
    CA
    95192, United States
    Int. J. Secur. Netw., 2 (124-136):
  • [25] A Compression-Based Method for Stemmatic Analysis
    Roos, Teemu
    Heikkila, Tuomas
    Myllymaki, Petri
    ECAI 2006, PROCEEDINGS, 2006, 141 : 805 - +
  • [26] Universal Lossless Compression-based Denoising
    Su, Han-I
    Weissman, Tsachy
    2010 IEEE INTERNATIONAL SYMPOSIUM ON INFORMATION THEORY, 2010, : 1648 - 1652
  • [27] Compression-based SoC Test Infrastructures
    Dalmasso, Julien
    Flottes, Marie-Lise
    Rouzeyre, Bruno
    VLSI-SOC: ADVANCED TOPICS ON SYSTEMS ON A CHIP, 2009, 291 : 53 - 67
  • [28] A compression-based text steganography method
    Satir, Esra
    Isik, Hakan
    JOURNAL OF SYSTEMS AND SOFTWARE, 2012, 85 (10) : 2385 - 2394
  • [29] Multilinear Multitask Learning by Rank-Product Regularization
    Zhao, Qian
    Rui, Xiangyu
    Han, Zhi
    Meng, Deyu
    IEEE TRANSACTIONS ON NEURAL NETWORKS AND LEARNING SYSTEMS, 2020, 31 (04) : 1336 - 1350
  • [30] Broad Multitask Learning System With Group Sparse Regularization
    Huang, Jintao
    Chen, Chuangquan
    Vong, Chi-Man
    Cheung, Yiu-Ming
    IEEE TRANSACTIONS ON NEURAL NETWORKS AND LEARNING SYSTEMS, 2024,