Large-Scale Fuzzy Least Squares Twin SVMs for Class Imbalance Learning

被引:31
|
作者
Ganaie, M. A. [1 ]
Tanveer, M. [1 ]
Lin, Chin-Teng [2 ]
机构
[1] Indian Inst Technol Indore, Dept Math, Indore 453552, India
[2] Univ Technol Sydney, Fac Engn & Informat Technol, Ctr Artificial Intelligence, Ultimo, NSW 2007, Australia
基金
美国国家卫生研究院;
关键词
Support vector machines; Kernel; Risk management; Minimization; Computational modeling; Alzheimer's disease; Data models; Alzheimer's disease (AD); class imbalance; machine learning; magnetic resonance imaging (MRI); maximum margin; mild cognitive impairment (MCI); pattern classification; structural risk minimization (SRM) principle; support vector machines (SVMs); twin support vector machine (TSVM); SUPPORT VECTOR MACHINES; CLASSIFICATION;
D O I
10.1109/TFUZZ.2022.3161729
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Twin support vector machines (TSVMs) have been successfully employed for binary classification problems. With the advent of machine learning algorithms, data have proliferated and there is a need to handle or process large-scale data. TSVMs are not successful in handling large-scale data due to the following: 1) the optimization problem solved in the TSVM needs to calculate large matrix inverses, which makes it an ineffective choice for large-scale problems; 2) the empirical risk minimization principle is employed in the TSVM and, hence, may suffer due to overfitting; and 3) the Wolfe dual of TSVM formulation involves positive-semidefinite matrices, and hence, singularity issues need to be resolved manually. Keeping in view the aforementioned shortcomings, in this article, we propose a novel large-scale fuzzy least squares TSVM for class imbalance learning (LS-FLSTSVM-CIL). We formulate the LS-FLSTSVM-CIL such that the proposed optimization problem ensures that: 1) no matrix inversion is involved in the proposed LS-FLSTSVM-CIL formulation, which makes it an efficient choice for large-scale problems; 2) the structural risk minimization principle is implemented, which avoids the issues of overfitting and results in better performance; and 3) the Wolfe dual formulation of the proposed LS-FLSTSVM-CIL model involves positive-definite matrices. In addition, to resolve the issues of class imbalance, we assign fuzzy weights in the proposed LS-FLSTSVM-CIL to avoid bias in dominating the samples of class imbalance problems. To make it more feasible for large-scale problems, we use an iterative procedure known as the sequential minimization principle to solve the objective function of the proposed LS-FLSTSVM-CIL model. From the experimental results, one can see that the proposed LS-FLSTSVM-CIL demonstrates superior performance in comparison to baseline classifiers. To demonstrate the feasibility of the proposed LS-FLSTSVM-CIL on large-scale classification problems, we evaluate the classification models on the large-scale normally distributed clustered (NDC) dataset. To demonstrate the practical applications of the proposed LS-FLSTSVM-CIL model, we evaluate it for the diagnosis of Alzheimer's disease and breast cancer disease. Evaluation on NDC datasets shows that the proposed LS-FLSTSVM-CIL has feasibility in large-scale problems as it is fast in comparison to the baseline classifiers.
引用
收藏
页码:4815 / 4827
页数:13
相关论文
共 50 条
  • [41] NONLINEAR LEAST-SQUARES APPROACH FOR LARGE-SCALE ALGEBRAIC RICCATI EQUATIONS
    Jbilou, Khalide
    Raydan, Marcos
    SIAM JOURNAL ON SCIENTIFIC COMPUTING, 2019, 41 (04): : A2193 - A2211
  • [42] BLOCK AOR ITERATIVE SCHEMES FOR LARGE-SCALE LEAST-SQUARES PROBLEMS
    PAPADOPOULOU, EP
    SARIDAKIS, YG
    PAPATHEODOROU, TS
    SIAM JOURNAL ON NUMERICAL ANALYSIS, 1989, 26 (03) : 637 - 660
  • [43] Quasi-Newton algorithms for large-scale nonlinear least-squares
    Al-Baali, M
    HIGH PERFORMANCE ALGORITHMS AND SOFTWARE FOR NONLINEAR OPTIMIZATION, 2003, 82 : 1 - 21
  • [44] SOLVING LARGE-SCALE LEAST SQUARES SEMIDEFINITE PROGRAMMING BY ALTERNATING DIRECTION METHODS
    He, Bingsheng
    Xu, Minghua
    Yuan, Xiaoming
    SIAM JOURNAL ON MATRIX ANALYSIS AND APPLICATIONS, 2011, 32 (01) : 136 - 152
  • [45] Distributed Least-Squares Iterative Methods in Large-Scale Networks:A Survey
    SHI Lei
    ZHAO Liang
    SONG Wenzhan
    Goutham Kamath
    WU Yuan
    LIU Xuefeng
    ZTECommunications, 2017, 15 (03) : 37 - 45
  • [46] Approximating sparse Hessian matrices using large-scale linear least squares
    Fowkes, Jaroslav M.
    Gould, Nicholas I. M.
    Scott, Jennifer A.
    NUMERICAL ALGORITHMS, 2024, 96 (04) : 1675 - 1698
  • [47] LARGE-SCALE GEODETIC LEAST-SQUARES ADJUSTMENT BY DISSECTION AND ORTHOGONAL DECOMPOSITION
    GOLUB, GH
    PLEMMONS, RJ
    LINEAR ALGEBRA AND ITS APPLICATIONS, 1980, 34 (DEC) : 3 - 28
  • [48] Gradient Projection Iterative Sketch for Large-Scale Constrained Least-Squares
    Tang, Junqi
    Golbabaee, Mohammad
    Davies, Mike E.
    INTERNATIONAL CONFERENCE ON MACHINE LEARNING, VOL 70, 2017, 70
  • [49] Expediting the Accuracy-Improving Process of SVMs for Class Imbalance Learning
    Cao, Bin
    Liu, Yuqi
    Hou, Chenyu
    Fan, Jing
    Zheng, Baihua
    Yin, Jianwei
    IEEE TRANSACTIONS ON KNOWLEDGE AND DATA ENGINEERING, 2021, 33 (11) : 3550 - 3567
  • [50] A fuzzy twin support vector machine based on information entropy for class imbalance learning
    Gupta, Deepak
    Richhariya, Bharat
    Borah, Parashjyoti
    NEURAL COMPUTING & APPLICATIONS, 2019, 31 (11): : 7153 - 7164