Fast Dictionary Learning with a Smoothed Wasserstein Loss

被引:0
|
作者
Rolet, Antoine [1 ]
Cuturi, Marco [1 ]
Peyre, Gabriel [2 ]
机构
[1] Kyoto Univ, Grad Sch Informat, Kyoto, Japan
[2] Univ Paris 09, CNRS, CEREMADE, Paris, France
关键词
MATRIX FACTORIZATION; EQUIVALENCE;
D O I
暂无
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
We consider in this paper the dictionary learning problem when the observations are normalized histograms of features. This problem can be tackled using non-negative matrix factorization approaches, using typically Euclidean or Kullback-Leibler fitting errors. Because these fitting errors are separable and treat each feature on equal footing, they are blind to any similarity the features may share. We assume in this work that we have prior knowledge on these features. To leverage this side-information, we propose to use the Wasserstein (a.k.a. earth mover's or optimal transport) distance as the fitting error between each original point and its reconstruction, and we propose scalable algorithms to to so. Our methods build upon Fenchel duality and entropic regularization of Wasserstein distances, which improves not only speed but also computational stability. We apply these techniques on face images and text documents. We show in particular that we can learn dictionaries (topics) for bag-of-word representations of texts using words that may not have appeared in the original texts, or even words that come from a different language than that used in the texts.
引用
收藏
页码:630 / 638
页数:9
相关论文
共 50 条
  • [1] Learning with a Wasserstein Loss
    Frogner, Charlie
    Zhang, Chiyuan
    Mobahi, Hossein
    Araya-Polo, Mauricio
    Poggio, Tomaso
    ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 28 (NIPS 2015), 2015, 28
  • [2] Geometrically Regularized Wasserstein Dictionary Learning
    Mueller, Marshall
    Aeron, Shuchin
    Murphy, James M.
    Tasissa, Abiy
    TOPOLOGICAL, ALGEBRAIC AND GEOMETRIC LEARNING WORKSHOPS 2023, VOL 221, 2023, 221
  • [3] Wasserstein of Wasserstein Loss for Learning Generative Models
    Dukler, Yonatan
    Li, Wuchen
    Lin, Alex Tong
    Montufar, Guido
    INTERNATIONAL CONFERENCE ON MACHINE LEARNING, VOL 97, 2019, 97
  • [4] Wasserstein Discriminant Dictionary Learning for Graph Representation
    Zhang, Tong
    Liu, Guangbu
    Cui, Zhen
    Liu, Wei
    Zheng, Wenming
    Yang, Jian
    IEEE TRANSACTIONS ON PATTERN ANALYSIS AND MACHINE INTELLIGENCE, 2024, 46 (12) : 8619 - 8635
  • [5] Wasserstein Discriminant Dictionary Learning for Graph Representation
    Zhang, Tong
    Liu, Guangbu
    Cui, Zhen
    Liu, Wei
    Zheng, Wenming
    Yang, Jian
    IEEE Transactions on Pattern Analysis and Machine Intelligence, 2024, 46 (12): : 8619 - 8635
  • [6] Wasserstein Dictionary Learning: Optimal Transport-Based Unsupervised Nonlinear Dictionary Learning
    Schmitz, Morgan A.
    Heitz, Matthieu
    Bonneel, Nicolas
    Ngole, Fred
    Coeurjolly, David
    Cuturi, Marco
    Peyre, Gabriel
    Starck, Jean-Luc
    SIAM JOURNAL ON IMAGING SCIENCES, 2018, 11 (01): : 643 - 678
  • [7] Asymptotics of Smoothed Wasserstein Distances
    Chen, Hong-Bin
    Niles-Weed, Jonathan
    POTENTIAL ANALYSIS, 2022, 56 (04) : 571 - 595
  • [8] Asymptotics of Smoothed Wasserstein Distances
    Hong-Bin Chen
    Jonathan Niles-Weed
    Potential Analysis, 2022, 56 : 571 - 595
  • [9] Multi-Level Metric Learning via Smoothed Wasserstein Distance
    Xu, Jie
    Luo, Lei
    Deng, Cheng
    Huang, Heng
    PROCEEDINGS OF THE TWENTY-SEVENTH INTERNATIONAL JOINT CONFERENCE ON ARTIFICIAL INTELLIGENCE, 2018, : 2919 - 2925
  • [10] SMOOTHED SIMCO FOR DICTIONARY LEARNING: HANDLING THE SINGULARITY ISSUE
    Zhao, Xiaochen
    Zhou, Guangyu
    Dai, Wei
    2013 IEEE INTERNATIONAL CONFERENCE ON ACOUSTICS, SPEECH AND SIGNAL PROCESSING (ICASSP), 2013, : 3292 - 3296