Large-Scale Fuzzy Least Squares Twin SVMs for Class Imbalance Learning

被引:31
|
作者
Ganaie, M. A. [1 ]
Tanveer, M. [1 ]
Lin, Chin-Teng [2 ]
机构
[1] Indian Inst Technol Indore, Dept Math, Indore 453552, India
[2] Univ Technol Sydney, Fac Engn & Informat Technol, Ctr Artificial Intelligence, Ultimo, NSW 2007, Australia
基金
美国国家卫生研究院;
关键词
Support vector machines; Kernel; Risk management; Minimization; Computational modeling; Alzheimer's disease; Data models; Alzheimer's disease (AD); class imbalance; machine learning; magnetic resonance imaging (MRI); maximum margin; mild cognitive impairment (MCI); pattern classification; structural risk minimization (SRM) principle; support vector machines (SVMs); twin support vector machine (TSVM); SUPPORT VECTOR MACHINES; CLASSIFICATION;
D O I
10.1109/TFUZZ.2022.3161729
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Twin support vector machines (TSVMs) have been successfully employed for binary classification problems. With the advent of machine learning algorithms, data have proliferated and there is a need to handle or process large-scale data. TSVMs are not successful in handling large-scale data due to the following: 1) the optimization problem solved in the TSVM needs to calculate large matrix inverses, which makes it an ineffective choice for large-scale problems; 2) the empirical risk minimization principle is employed in the TSVM and, hence, may suffer due to overfitting; and 3) the Wolfe dual of TSVM formulation involves positive-semidefinite matrices, and hence, singularity issues need to be resolved manually. Keeping in view the aforementioned shortcomings, in this article, we propose a novel large-scale fuzzy least squares TSVM for class imbalance learning (LS-FLSTSVM-CIL). We formulate the LS-FLSTSVM-CIL such that the proposed optimization problem ensures that: 1) no matrix inversion is involved in the proposed LS-FLSTSVM-CIL formulation, which makes it an efficient choice for large-scale problems; 2) the structural risk minimization principle is implemented, which avoids the issues of overfitting and results in better performance; and 3) the Wolfe dual formulation of the proposed LS-FLSTSVM-CIL model involves positive-definite matrices. In addition, to resolve the issues of class imbalance, we assign fuzzy weights in the proposed LS-FLSTSVM-CIL to avoid bias in dominating the samples of class imbalance problems. To make it more feasible for large-scale problems, we use an iterative procedure known as the sequential minimization principle to solve the objective function of the proposed LS-FLSTSVM-CIL model. From the experimental results, one can see that the proposed LS-FLSTSVM-CIL demonstrates superior performance in comparison to baseline classifiers. To demonstrate the feasibility of the proposed LS-FLSTSVM-CIL on large-scale classification problems, we evaluate the classification models on the large-scale normally distributed clustered (NDC) dataset. To demonstrate the practical applications of the proposed LS-FLSTSVM-CIL model, we evaluate it for the diagnosis of Alzheimer's disease and breast cancer disease. Evaluation on NDC datasets shows that the proposed LS-FLSTSVM-CIL has feasibility in large-scale problems as it is fast in comparison to the baseline classifiers.
引用
收藏
页码:4815 / 4827
页数:13
相关论文
共 50 条
  • [11] LARGE-SCALE DUAL REGULARIZED TOTAL LEAST SQUARES
    Lampe, Joerg
    Voss, Heinrich
    ELECTRONIC TRANSACTIONS ON NUMERICAL ANALYSIS, 2014, 42 : 13 - 40
  • [12] Large-scale Tikhonov regularization of total least squares
    Lampe, Joerg
    Voss, Heinrich
    JOURNAL OF COMPUTATIONAL AND APPLIED MATHEMATICS, 2012, 238 : 95 - 108
  • [13] Model reduction of large-scale systems by least squares
    Gugercin, Serkan
    Antoulas, Athanasios C.
    LINEAR ALGEBRA AND ITS APPLICATIONS, 2006, 415 (2-3) : 290 - 321
  • [14] On solving large-scale weighted least squares problems
    Baryamureeba, V
    NUMERICAL ANALYSIS AND ITS APPLICATIONS, 2001, 1988 : 59 - 67
  • [15] Core-elements for large-scale least squares estimation
    Li, Mengyu
    Yu, Jun
    Li, Tao
    Meng, Cheng
    STATISTICS AND COMPUTING, 2024, 34 (06)
  • [16] A Universal Analysis of Large-Scale Regularized Least Squares Solutions
    Panahi, Ashkan
    Hassibi, Babak
    ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 30 (NIPS 2017), 2017, 30
  • [17] Solving large-scale constrained least-squares problems
    Abdel-Aziz, MR
    El-Alem, MM
    APPLIED MATHEMATICS AND COMPUTATION, 2003, 137 (2-3) : 571 - 587
  • [18] Large-Scale Regression: A Partition Analysis of the Least Squares Multisplitting
    Inghelbrecht, Gilles
    Pintelon, Rik
    Barbe, Kurt
    IEEE TRANSACTIONS ON INSTRUMENTATION AND MEASUREMENT, 2020, 69 (06) : 2635 - 2647
  • [19] Solution of large-scale weighted least-squares problems
    Baryamureeba, V
    NUMERICAL LINEAR ALGEBRA WITH APPLICATIONS, 2002, 9 (02) : 93 - 106
  • [20] Scaled Least Squares Estimator for GLMs in Large-Scale Problems
    Erdogdu, Murat A.
    Bayati, Mohsen
    Dicker, Lee H.
    ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 29 (NIPS 2016), 2016, 29