Large-Scale Fuzzy Least Squares Twin SVMs for Class Imbalance Learning

被引:31
|
作者
Ganaie, M. A. [1 ]
Tanveer, M. [1 ]
Lin, Chin-Teng [2 ]
机构
[1] Indian Inst Technol Indore, Dept Math, Indore 453552, India
[2] Univ Technol Sydney, Fac Engn & Informat Technol, Ctr Artificial Intelligence, Ultimo, NSW 2007, Australia
基金
美国国家卫生研究院;
关键词
Support vector machines; Kernel; Risk management; Minimization; Computational modeling; Alzheimer's disease; Data models; Alzheimer's disease (AD); class imbalance; machine learning; magnetic resonance imaging (MRI); maximum margin; mild cognitive impairment (MCI); pattern classification; structural risk minimization (SRM) principle; support vector machines (SVMs); twin support vector machine (TSVM); SUPPORT VECTOR MACHINES; CLASSIFICATION;
D O I
10.1109/TFUZZ.2022.3161729
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Twin support vector machines (TSVMs) have been successfully employed for binary classification problems. With the advent of machine learning algorithms, data have proliferated and there is a need to handle or process large-scale data. TSVMs are not successful in handling large-scale data due to the following: 1) the optimization problem solved in the TSVM needs to calculate large matrix inverses, which makes it an ineffective choice for large-scale problems; 2) the empirical risk minimization principle is employed in the TSVM and, hence, may suffer due to overfitting; and 3) the Wolfe dual of TSVM formulation involves positive-semidefinite matrices, and hence, singularity issues need to be resolved manually. Keeping in view the aforementioned shortcomings, in this article, we propose a novel large-scale fuzzy least squares TSVM for class imbalance learning (LS-FLSTSVM-CIL). We formulate the LS-FLSTSVM-CIL such that the proposed optimization problem ensures that: 1) no matrix inversion is involved in the proposed LS-FLSTSVM-CIL formulation, which makes it an efficient choice for large-scale problems; 2) the structural risk minimization principle is implemented, which avoids the issues of overfitting and results in better performance; and 3) the Wolfe dual formulation of the proposed LS-FLSTSVM-CIL model involves positive-definite matrices. In addition, to resolve the issues of class imbalance, we assign fuzzy weights in the proposed LS-FLSTSVM-CIL to avoid bias in dominating the samples of class imbalance problems. To make it more feasible for large-scale problems, we use an iterative procedure known as the sequential minimization principle to solve the objective function of the proposed LS-FLSTSVM-CIL model. From the experimental results, one can see that the proposed LS-FLSTSVM-CIL demonstrates superior performance in comparison to baseline classifiers. To demonstrate the feasibility of the proposed LS-FLSTSVM-CIL on large-scale classification problems, we evaluate the classification models on the large-scale normally distributed clustered (NDC) dataset. To demonstrate the practical applications of the proposed LS-FLSTSVM-CIL model, we evaluate it for the diagnosis of Alzheimer's disease and breast cancer disease. Evaluation on NDC datasets shows that the proposed LS-FLSTSVM-CIL has feasibility in large-scale problems as it is fast in comparison to the baseline classifiers.
引用
收藏
页码:4815 / 4827
页数:13
相关论文
共 50 条
  • [21] M-Decomposed Least Squares and Recursive Least Squares Identification Algorithms for Large-Scale Systems
    Ji, Yuejiang
    Lv, Lixin
    IEEE ACCESS, 2021, 9 : 139466 - 139472
  • [22] Large-scale linear nonparallel SVMs
    Liu, Dalian
    Li, Dewei
    Shi, Yong
    Tian, Yingjie
    SOFT COMPUTING, 2018, 22 (06) : 1945 - 1957
  • [23] Large-scale linear nonparallel SVMs
    Dalian Liu
    Dewei Li
    Yong Shi
    Yingjie Tian
    Soft Computing, 2018, 22 : 1945 - 1957
  • [24] Large-scale classification by an Approximate Least Squares One-Class Support Vector Machine ensemble
    Mygdalis, Vasileios
    Iosifidis, Alexandros
    Tefas, Anastasios
    Pitas, Ioannis
    2015 IEEE TRUSTCOM/BIGDATASE/ISPA, VOL 2, 2015, : 6 - 10
  • [25] Accelerated orthogonal least-squares for large-scale sparse reconstruction
    Hashemi, Abolfazl
    Vikalo, Haris
    DIGITAL SIGNAL PROCESSING, 2018, 82 : 91 - 105
  • [26] Partitioned least-squares operator for large-scale geophysical inversion
    Porsani, Milton J.
    Stoffa, Paul L.
    Sen, Mrinal K.
    Seif, Roustam K.
    GEOPHYSICS, 2010, 75 (06) : R121 - R128
  • [27] Constrained Stochastic Gradient Descent for Large-scale Least Squares Problem
    Mu, Yang
    Ding, Wei
    Zhou, Tianyi
    Tao, Dacheng
    19TH ACM SIGKDD INTERNATIONAL CONFERENCE ON KNOWLEDGE DISCOVERY AND DATA MINING (KDD'13), 2013, : 883 - 891
  • [28] INCREMENTAL REGULARIZED LEAST SQUARES FOR DIMENSIONALITY REDUCTION OF LARGE-SCALE DATA
    Zhang, Xiaowei
    Cheng, Li
    Chu, Delin
    Liao, Li-Zhi
    Ng, Michael K.
    Tan, Roger C. E.
    SIAM JOURNAL ON SCIENTIFIC COMPUTING, 2016, 38 (03): : B414 - B439
  • [29] Analysis of extended partial least squares for monitoring large-scale processes
    Chen, Q
    Kruger, U
    IEEE TRANSACTIONS ON CONTROL SYSTEMS TECHNOLOGY, 2005, 13 (05) : 807 - 813
  • [30] Fuzzy least squares twin support vector clustering
    Reshma Khemchandani
    Aman Pal
    Suresh Chandra
    Neural Computing and Applications, 2018, 29 : 553 - 563