Convergence of Non-Convex Non-Concave GANs Using Sinkhorn Divergence

被引:1
|
作者
Adnan, Risman [1 ,2 ]
Saputra, Muchlisin Adi [1 ]
Fadlil, Junaidillah [1 ]
Ezerman, Martianus Frederic [3 ]
Iqbal, Muhamad [2 ]
Basaruddin, Tjan [2 ]
机构
[1] Samsung R&D Indonesia SRIN, Jakarta 10210, Indonesia
[2] Univ Indonesia, Dept Comp Sci, Depok 16424, Indonesia
[3] Nanyang Technol Univ, Sch Phys & Math Sci, Singapore 637371, Singapore
关键词
Convergence; generative adversarial networks; optimal transport; Sinkhorn divergence;
D O I
10.1109/ACCESS.2021.3074943
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
Sinkhorn divergence is a symmetric normalization of entropic regularized optimal transport. It is a smooth and continuous metrized weak-convergence with excellent geometric properties. We use it as an alternative for the minimax objective function in formulating generative adversarial networks. The optimization is defined with Sinkhorn divergence as the objective, under the non-convex and non-concave condition. This work focuses on the optimization's convergence and stability. We propose a first order sequential stochastic gradient descent ascent (SeqSGDA) algorithm. Under some mild approximations, the learning converges to local minimax points. Using the structural similarity index measure (SSIM), we supply a non-asymptotic analysis of the algorithm's convergence rate. Empirical evidences show a convergence rate, which is inversely proportional to the number of iterations, when tested on tiny colour datasets Cats and CelebA on the deep convolutional generative adversarial networks and ResNet neural architectures. The entropy regularization parameters is approximated to the SSIM tolerance epsilon. We determine that the iteration complexity to return to an epsilon-stationary point to be O (kappa log(epsilon(-1))), where kappa is a value that depends on the Sinkhorn divergence's convexity and the minimax step ratio in the SeqSGDA algorithm.
引用
收藏
页码:67595 / 67609
页数:15
相关论文
共 50 条
  • [1] Improved Portfolio Optimization with Non-convex and Non-concave Cost Using Genetic Algorithms
    Lu, Zhang
    Wang, Xiaoli
    PROCEEDINGS 2013 INTERNATIONAL CONFERENCE ON MECHATRONIC SCIENCES, ELECTRIC ENGINEERING AND COMPUTER (MEC), 2013, : 2567 - 2570
  • [2] A decentralized adaptive method with consensus step for non-convex non-concave min-max optimization problems
    Li, Meiwen
    Long, Xinyue
    Liu, Muhua
    Guo, Jing
    Zhao, Xuhui
    Wang, Lin
    Wu, Qingtao
    EXPERT SYSTEMS WITH APPLICATIONS, 2025, 276
  • [3] Non-concave dynamic programming
    Cotter, KD
    Park, JH
    ECONOMICS LETTERS, 2006, 90 (01) : 141 - 146
  • [4] Poincare Recurrence, Cycles and Spurious Equilibria in Gradient-Descent-Ascent for Non-Convex Non-Concave Zero-Sum Games
    Flokas, Lampros
    Vlatakis-Gkaragkounis, Emmanouil V.
    Piliouras, Georgios
    ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 32 (NIPS 2019), 2019, 32
  • [5] MINIMIZATION OF CONVEX FUNCTIONALS INVOLVING NESTED MAXIMA - NON-CONCAVE DUALITY AND ALGORITHMS
    COHEN, G
    JOURNAL OF OPTIMIZATION THEORY AND APPLICATIONS, 1982, 36 (03) : 335 - 365
  • [6] Nesting of non-convex figures in non-convex contours
    Vinade, C.
    Dias, A.
    Informacion Tecnologica, 2000, 11 (01): : 149 - 156
  • [7] Γ-convergence of non-local, non-convex functionals in one dimension
    Brezis, Haim
    Hoai-Minh Nguyen
    COMMUNICATIONS IN CONTEMPORARY MATHEMATICS, 2020, 22 (07)
  • [8] Global convergence of a curvilinear search for non-convex optimization
    Bartholomew-Biggs, Michael
    Beddiaf, Salah
    Christianson, Bruce
    NUMERICAL ALGORITHMS, 2023, 92 (04) : 2025 - 2043
  • [9] Uniform Convergence of Gradients for Non-Convex Learning and Optimization
    Foster, Dylan J.
    Sekhari, Ayush
    Sridharan, Karthik
    ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 31 (NIPS 2018), 2018, 31
  • [10] The Convergence Guarantees of a Non-Convex Approach for Sparse Recovery
    Chen, Laming
    Gu, Yuantao
    IEEE TRANSACTIONS ON SIGNAL PROCESSING, 2014, 62 (15) : 3754 - 3767