Learning with a Wasserstein Loss

被引:0
|
作者
Frogner, Charlie [1 ]
Zhang, Chiyuan [1 ]
Mobahi, Hossein [2 ]
Araya-Polo, Mauricio [3 ]
Poggio, Tomaso [1 ]
机构
[1] MIT, Ctr Brains Minds & Machines, Cambridge, MA 02139 USA
[2] MIT, CSAIL, Cambridge, MA 02139 USA
[3] Shell Int E&P Inc, The Hague, Netherlands
关键词
DISTANCE;
D O I
暂无
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Learning to predict multi-label outputs is challenging, but in many problems there is a natural metric on the outputs that can be used to improve predictions. In this paper we develop a loss function for multi-label learning, based on the Wasserstein distance. The Wasserstein distance provides a natural notion of dissimilarity for probability measures. Although optimizing with respect to the exact Wasserstein distance is costly, recent work has described a regularized approximation that is efficiently computed. We describe an efficient learning algorithm based on this regularization, as well as a novel extension of the Wasserstein distance from probability measures to unnormalized measures. We also describe a statistical learning bound for the loss. The Wasserstein loss can encourage smoothness of the predictions with respect to a chosen metric on the output space. We demonstrate this property on a real-data tag prediction problem, using the Yahoo Flickr Creative Commons dataset, outperforming a baseline that doesn't use the metric.
引用
收藏
页数:9
相关论文
共 50 条
  • [21] WRGAN: Improvement of RelGAN with Wasserstein Loss for Text Generation
    Jiao, Ziyun
    Ren, Fuji
    ELECTRONICS, 2021, 10 (03) : 1 - 14
  • [22] Deep Distributional Sequence Embeddings Based on a Wasserstein Loss
    Ahmed Abdelwahab
    Niels Landwehr
    Neural Processing Letters, 2022, 54 : 3749 - 3769
  • [23] Deep Distributional Sequence Embeddings Based on a Wasserstein Loss
    Abdelwahab, Ahmed
    Landwehr, Niels
    NEURAL PROCESSING LETTERS, 2022, 54 (05) : 3749 - 3769
  • [24] GeONet: a neural operator for learning the Wasserstein geodesic
    Gracyk, Andrew
    Chen, Xiaohui
    UNCERTAINTY IN ARTIFICIAL INTELLIGENCE, 2024, 244 : 1453 - 1478
  • [25] Wasserstein Discriminant Dictionary Learning for Graph Representation
    Zhang, Tong
    Liu, Guangbu
    Cui, Zhen
    Liu, Wei
    Zheng, Wenming
    Yang, Jian
    IEEE TRANSACTIONS ON PATTERN ANALYSIS AND MACHINE INTELLIGENCE, 2024, 46 (12) : 8619 - 8635
  • [26] Learning with minibatch Wasserstein : asymptotic and gradient properties
    Fatras, Kilian
    Zine, Younes
    Flamary, Remi
    Gribonval, Remi
    Courty, Nicolas
    INTERNATIONAL CONFERENCE ON ARTIFICIAL INTELLIGENCE AND STATISTICS, VOL 108, 2020, 108 : 2131 - 2140
  • [27] Wasserstein Propagation for Semi-Supervised Learning
    Solomon, Justin
    Rustamov, Raif M.
    Guibas, Leonidas
    Butscher, Adrian
    INTERNATIONAL CONFERENCE ON MACHINE LEARNING, VOL 32 (CYCLE 1), 2014, 32
  • [28] Wasserstein Adversarial Regularization for Learning With Label Noise
    Fatras, Kilian
    Damodaran, Bharath Bhushan
    Lobry, Sylvain
    Flamary, Remi
    Tuia, Devis
    Courty, Nicolas
    IEEE TRANSACTIONS ON PATTERN ANALYSIS AND MACHINE INTELLIGENCE, 2022, 44 (10) : 7296 - 7306
  • [29] Learning Wasserstein Isometric Embedding for Point Clouds
    Kawano, Keisuke
    Koide, Satoshi
    Kutsuna, Takuro
    2020 INTERNATIONAL CONFERENCE ON 3D VISION (3DV 2020), 2020, : 473 - 482
  • [30] Wasserstein Discriminant Dictionary Learning for Graph Representation
    Zhang, Tong
    Liu, Guangbu
    Cui, Zhen
    Liu, Wei
    Zheng, Wenming
    Yang, Jian
    IEEE Transactions on Pattern Analysis and Machine Intelligence, 2024, 46 (12): : 8619 - 8635