Large-Scale Fuzzy Least Squares Twin SVMs for Class Imbalance Learning

被引:31
|
作者
Ganaie, M. A. [1 ]
Tanveer, M. [1 ]
Lin, Chin-Teng [2 ]
机构
[1] Indian Inst Technol Indore, Dept Math, Indore 453552, India
[2] Univ Technol Sydney, Fac Engn & Informat Technol, Ctr Artificial Intelligence, Ultimo, NSW 2007, Australia
基金
美国国家卫生研究院;
关键词
Support vector machines; Kernel; Risk management; Minimization; Computational modeling; Alzheimer's disease; Data models; Alzheimer's disease (AD); class imbalance; machine learning; magnetic resonance imaging (MRI); maximum margin; mild cognitive impairment (MCI); pattern classification; structural risk minimization (SRM) principle; support vector machines (SVMs); twin support vector machine (TSVM); SUPPORT VECTOR MACHINES; CLASSIFICATION;
D O I
10.1109/TFUZZ.2022.3161729
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Twin support vector machines (TSVMs) have been successfully employed for binary classification problems. With the advent of machine learning algorithms, data have proliferated and there is a need to handle or process large-scale data. TSVMs are not successful in handling large-scale data due to the following: 1) the optimization problem solved in the TSVM needs to calculate large matrix inverses, which makes it an ineffective choice for large-scale problems; 2) the empirical risk minimization principle is employed in the TSVM and, hence, may suffer due to overfitting; and 3) the Wolfe dual of TSVM formulation involves positive-semidefinite matrices, and hence, singularity issues need to be resolved manually. Keeping in view the aforementioned shortcomings, in this article, we propose a novel large-scale fuzzy least squares TSVM for class imbalance learning (LS-FLSTSVM-CIL). We formulate the LS-FLSTSVM-CIL such that the proposed optimization problem ensures that: 1) no matrix inversion is involved in the proposed LS-FLSTSVM-CIL formulation, which makes it an efficient choice for large-scale problems; 2) the structural risk minimization principle is implemented, which avoids the issues of overfitting and results in better performance; and 3) the Wolfe dual formulation of the proposed LS-FLSTSVM-CIL model involves positive-definite matrices. In addition, to resolve the issues of class imbalance, we assign fuzzy weights in the proposed LS-FLSTSVM-CIL to avoid bias in dominating the samples of class imbalance problems. To make it more feasible for large-scale problems, we use an iterative procedure known as the sequential minimization principle to solve the objective function of the proposed LS-FLSTSVM-CIL model. From the experimental results, one can see that the proposed LS-FLSTSVM-CIL demonstrates superior performance in comparison to baseline classifiers. To demonstrate the feasibility of the proposed LS-FLSTSVM-CIL on large-scale classification problems, we evaluate the classification models on the large-scale normally distributed clustered (NDC) dataset. To demonstrate the practical applications of the proposed LS-FLSTSVM-CIL model, we evaluate it for the diagnosis of Alzheimer's disease and breast cancer disease. Evaluation on NDC datasets shows that the proposed LS-FLSTSVM-CIL has feasibility in large-scale problems as it is fast in comparison to the baseline classifiers.
引用
收藏
页码:4815 / 4827
页数:13
相关论文
共 50 条
  • [31] Fuzzy least squares twin support vector machines
    Sartakhti, Javad Salimi
    Afrabandpey, Homayun
    Ghadiri, Nasser
    ENGINEERING APPLICATIONS OF ARTIFICIAL INTELLIGENCE, 2019, 85 : 402 - 409
  • [32] Fuzzy least squares twin support vector clustering
    Khemchandani, Reshma
    Pal, Aman
    Chandra, Suresh
    NEURAL COMPUTING & APPLICATIONS, 2018, 29 (02): : 553 - 563
  • [33] High-Performance Machine Learning for Large-Scale Data Classification considering Class Imbalance
    Liu, Yang
    Li, Xiang
    Chen, Xianbang
    Wang, Xi
    Li, Huaqiang
    SCIENTIFIC PROGRAMMING, 2020, 2020 (2020)
  • [34] DeepLM: Large-scale Nonlinear Least Squares on Deep Learning Frameworks using Stochastic Domain Decomposition
    Huang, Jingwei
    Huang, Shan
    Sun, Mingwei
    2021 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION, CVPR 2021, 2021, : 10303 - 10312
  • [35] Class imbalance learning using fuzzy ART and intuitionistic fuzzy twin support vector machines
    Rezvani, Salim
    Wang, Xizhao
    INFORMATION SCIENCES, 2021, 578 (578) : 659 - 682
  • [36] Fuzzy twin support vector machine based on affinity and class probability for class imbalance learning
    Hazarika, Barenya Bikash
    Gupta, Deepak
    Borah, Parashjyoti
    KNOWLEDGE AND INFORMATION SYSTEMS, 2023, 65 (12) : 5259 - 5288
  • [37] Fuzzy twin support vector machine based on affinity and class probability for class imbalance learning
    Barenya Bikash Hazarika
    Deepak Gupta
    Parashjyoti Borah
    Knowledge and Information Systems, 2023, 65 : 5259 - 5288
  • [38] Distributed Weighted Least Squares Estimation with Fast Convergence in Large-scale Systems
    Marelli, Damin
    Fu, Minyue
    2013 IEEE 52ND ANNUAL CONFERENCE ON DECISION AND CONTROL (CDC), 2013, : 5432 - 5437
  • [39] Least squares estimation of spatial autoregressive models for large-scale social networks
    Huang, Danyang
    Lan, Wei
    Zhang, Hao Helen
    Wang, Hansheng
    ELECTRONIC JOURNAL OF STATISTICS, 2019, 13 (01): : 1135 - 1165
  • [40] The canonical least squares estimation of large-scale simultaneous-equations models
    Kang, Heejoon
    ECONOMIC MODELLING, 2008, 25 (02) : 191 - 200