Learning with a Wasserstein Loss

被引:0
|
作者
Frogner, Charlie [1 ]
Zhang, Chiyuan [1 ]
Mobahi, Hossein [2 ]
Araya-Polo, Mauricio [3 ]
Poggio, Tomaso [1 ]
机构
[1] MIT, Ctr Brains Minds & Machines, Cambridge, MA 02139 USA
[2] MIT, CSAIL, Cambridge, MA 02139 USA
[3] Shell Int E&P Inc, The Hague, Netherlands
关键词
DISTANCE;
D O I
暂无
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Learning to predict multi-label outputs is challenging, but in many problems there is a natural metric on the outputs that can be used to improve predictions. In this paper we develop a loss function for multi-label learning, based on the Wasserstein distance. The Wasserstein distance provides a natural notion of dissimilarity for probability measures. Although optimizing with respect to the exact Wasserstein distance is costly, recent work has described a regularized approximation that is efficiently computed. We describe an efficient learning algorithm based on this regularization, as well as a novel extension of the Wasserstein distance from probability measures to unnormalized measures. We also describe a statistical learning bound for the loss. The Wasserstein loss can encourage smoothness of the predictions with respect to a chosen metric on the output space. We demonstrate this property on a real-data tag prediction problem, using the Yahoo Flickr Creative Commons dataset, outperforming a baseline that doesn't use the metric.
引用
收藏
页数:9
相关论文
共 50 条
  • [1] Wasserstein of Wasserstein Loss for Learning Generative Models
    Dukler, Yonatan
    Li, Wuchen
    Lin, Alex Tong
    Montufar, Guido
    INTERNATIONAL CONFERENCE ON MACHINE LEARNING, VOL 97, 2019, 97
  • [2] Fast Dictionary Learning with a Smoothed Wasserstein Loss
    Rolet, Antoine
    Cuturi, Marco
    Peyre, Gabriel
    ARTIFICIAL INTELLIGENCE AND STATISTICS, VOL 51, 2016, 51 : 630 - 638
  • [3] Wasserstein distance loss function for financial time series deep learning
    Souto, Hugo Gobato
    Moradi, Amir
    SOFTWARE IMPACTS, 2024, 20
  • [4] Wasserstein Loss With Alternative Reinforcement Learning for Severity-Aware Semantic Segmentation
    Liu, Xiaofeng
    Lu, Yunhong
    Liu, Xiongchang
    Bai, Song
    Li, Site
    You, Jane
    IEEE TRANSACTIONS ON INTELLIGENT TRANSPORTATION SYSTEMS, 2022, 23 (01) : 587 - 596
  • [5] Wasserstein Loss for Image Synthesis and Restoration
    Tartavel, Guillaume
    Peyre, Gabriel
    Gousseau, Yann
    SIAM JOURNAL ON IMAGING SCIENCES, 2016, 9 (04): : 1726 - 1755
  • [6] Learning to Generate Wasserstein Barycenters
    Lacombe, Julien
    Digne, Julie
    Courty, Nicolas
    Bonneel, Nicolas
    JOURNAL OF MATHEMATICAL IMAGING AND VISION, 2023, 65 (02) : 354 - 370
  • [7] Bayesian learning with Wasserstein barycenters*
    Backhoff-Veraguas, Julio
    Fontbona, Joaquin
    Rios, Gonzalo
    Tobar, Felipe
    ESAIM-PROBABILITY AND STATISTICS, 2022, 26 : 436 - 472
  • [8] Learning to Generate Wasserstein Barycenters
    Julien Lacombe
    Julie Digne
    Nicolas Courty
    Nicolas Bonneel
    Journal of Mathematical Imaging and Vision, 2023, 65 : 354 - 370
  • [9] Statistical Learning in Wasserstein Space
    Karimi, Amirhossein
    Ripani, Luigia
    Georgiou, Tryphon T.
    IEEE CONTROL SYSTEMS LETTERS, 2021, 5 (03): : 899 - 904
  • [10] Wasserstein Unsupervised Reinforcement Learning
    He, Shuncheng
    Jiang, Yuhang
    Zhang, Hongchang
    Shao, Jianzhun
    Ji, Xiangyang
    THIRTY-SIXTH AAAI CONFERENCE ON ARTIFICIAL INTELLIGENCE / THIRTY-FOURTH CONFERENCE ON INNOVATIVE APPLICATIONS OF ARTIFICIAL INTELLIGENCE / THE TWELVETH SYMPOSIUM ON EDUCATIONAL ADVANCES IN ARTIFICIAL INTELLIGENCE, 2022, : 6884 - 6892