Nonlinear Regression via Deep Negative Correlation Learning

被引:50
|
作者
Zhang, Le [1 ]
Shi, Zenglin [2 ]
Cheng, Ming-Ming [3 ]
Liu, Yun [3 ]
Bian, Jia-Wang [4 ]
Zhou, Joey Tianyi [1 ]
Zheng, Guoyan [5 ]
Zeng, Zeng [1 ]
机构
[1] ASTAR, Singapore 138632, Singapore
[2] Univ Amsterdam, NL-1012 WX Amsterdam, Netherlands
[3] Nankai Univ, Coll Comp Sci, TKLNDST, Nankai 300071, Peoples R China
[4] Univ Adelaide, Sch Comp Sci, Adelaide, SA 5005, Australia
[5] Shanghai Jiao Tong Univ, Sch Biomed Engn, Shanghai 200240, Peoples R China
关键词
Task analysis; Estimation; Training; Correlation; Computational modeling; Deep learning; Computer vision; deep regression; negative correlation learning; convolutional neural network; IMAGE QUALITY ASSESSMENT; HUMAN AGE ESTIMATION; SUPERRESOLUTION; ENSEMBLES; CASCADE;
D O I
10.1109/TPAMI.2019.2943860
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Nonlinear regression has been extensively employed in many computer vision problems (e.g., crowd counting, age estimation, affective computing). Under the umbrella of deep learning, two common solutions exist i) transforming nonlinear regression to a robust loss function which is jointly optimizable with the deep convolutional network, and ii) utilizing ensemble of deep networks. Although some improved performance is achieved, the former may be lacking due to the intrinsic limitation of choosing a single hypothesis and the latter may suffer from much larger computational complexity. To cope with those issues, we propose to regress via an efficient "divide and conquer" manner. The core of our approach is the generalization of negative correlation learning that has been shown, both theoretically and empirically, to work well for non-deep regression problems. Without extra parameters, the proposed method controls the bias-variance-covariance trade-off systematically and usually yields a deep regression ensemble where each base model is both "accurate" and "diversified." Moreover, we show that each sub-problem in the proposed method has less Rademacher Complexity and thus is easier to optimize. Extensive experiments on several diverse and challenging tasks including crowd counting, personality analysis, age estimation, and image super-resolution demonstrate the superiority over challenging baselines as well as the versatility of the proposed method. The source code and trained models are available on our project page: https://mmcheng.net/dncl/.
引用
收藏
页码:982 / 998
页数:17
相关论文
共 50 条
  • [41] Relieving Coefficient Learning in Genetic Programming for Symbolic Regression via Correlation and Linear Scaling
    Chen, Qi
    Xue, Bing
    Banzhaf, Wolfgang
    Zhang, Mengjie
    PROCEEDINGS OF THE 2023 GENETIC AND EVOLUTIONARY COMPUTATION CONFERENCE, GECCO 2023, 2023, : 420 - 437
  • [42] FAST APPROXIMATION OF NON-NEGATIVE SPARSE RECOVERY VIA DEEP LEARNING
    Xie, Youye
    Wang, Zifan
    Pei, Weiping
    Tang, Gongguo
    2019 IEEE INTERNATIONAL CONFERENCE ON IMAGE PROCESSING (ICIP), 2019, : 2921 - 2925
  • [43] Cross-Media Retrieval via Deep Semantic Canonical Correlation Analysis and Logistic Regression
    Zhang, Hong
    Xia, Liangmeng
    ADVANCES IN MULTIMEDIA INFORMATION PROCESSING, PT III, 2018, 11166 : 123 - 133
  • [44] A Nonlinear Regression Application via Machine Learning Techniques for Geomagnetic Data Reconstruction Processing
    Liu, Huan
    Liu, Zheng
    Liu, Shuo
    Liu, Yihao
    Bin, Junchi
    Shi, Fang
    Dong, Haobin
    IEEE TRANSACTIONS ON GEOSCIENCE AND REMOTE SENSING, 2019, 57 (01): : 128 - 140
  • [45] Prediction of emission and performance of internal combustion engine via regression deep learning approach
    Soltanalizadeh, Samaneh
    Yazdi, Mohammad Reza Haeri
    Esfahanian, Vahid
    Nejat, Mohammad
    PROCEEDINGS OF THE INSTITUTION OF MECHANICAL ENGINEERS PART D-JOURNAL OF AUTOMOBILE ENGINEERING, 2024,
  • [46] Deep learning baseline correction method via multi-scale analysis and regression
    Jiao, Qingliang
    Guo, Xiuwen
    Liu, Ming
    Kong, Lingqin
    Hui, Mei
    Dong, Liquan
    Zhao, Yuejin
    CHEMOMETRICS AND INTELLIGENT LABORATORY SYSTEMS, 2023, 235
  • [47] Regression Fuzzing for Deep Learning Systems
    You, Hanmo
    Wang, Zan
    Chen, Junjie
    Liu, Shuang
    Li, Shuochuan
    2023 IEEE/ACM 45TH INTERNATIONAL CONFERENCE ON SOFTWARE ENGINEERING, ICSE, 2023, : 82 - 94
  • [48] A deep learning approach to censored regression
    Danaila, Vlad-Rares
    Buiu, Catalin
    PATTERN ANALYSIS AND APPLICATIONS, 2024, 27 (01)
  • [49] Deep Metric Learning for Crowdedness Regression
    Wang, Qi
    Wan, Jia
    Yuan, Yuan
    IEEE TRANSACTIONS ON CIRCUITS AND SYSTEMS FOR VIDEO TECHNOLOGY, 2018, 28 (10) : 2633 - 2643
  • [50] SSIMLayer: Towards Robust Deep Representation Learning via Nonlinear Structural Similarity
    Abobakr, Ahmed
    Hossny, Mohammed
    Nahavandi, Saeid
    2019 IEEE INTERNATIONAL CONFERENCE ON SYSTEMS, MAN AND CYBERNETICS (SMC), 2019, : 1234 - 1238