Multimodal Representation Learning via Maximization of Local Mutual Information

被引:23
|
作者
Liao, Ruizhi [1 ]
Moyer, Daniel [1 ]
Cha, Miriam [2 ]
Quigley, Keegan [2 ]
Berkowitz, Seth [3 ]
Horng, Steven [3 ]
Golland, Polina [1 ]
Wells, William M. [1 ,4 ]
机构
[1] MIT, CSAIL, 77 Massachusetts Ave, Cambridge, MA 02139 USA
[2] MIT, Lincoln Lab, 244 Wood St, Lexington, MA 02173 USA
[3] Harvard Med Sch, Beth Israel Deaconess Med Ctr, Boston, MA 02115 USA
[4] Harvard Med Sch, Brigham & Womens Hosp, Boston, MA 02115 USA
关键词
Multimodal representation learning; Local feature representations; Mutual information maximization;
D O I
10.1007/978-3-030-87196-3_26
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
We propose and demonstrate a representation learning approach by maximizing the mutual information between local features of images and text. The goal of this approach is to learn useful image representations by taking advantage of the rich information contained in the free text that describes the findings in the image. Our method trains image and text encoders by encouraging the resulting representations to exhibit high local mutual information. We make use of recent advances in mutual information estimation with neural network discriminators. We argue that the sum of local mutual information is typically a lower bound on the global mutual information. Our experimental results in the downstream image classification tasks demonstrate the advantages of using local features for image-text representation learning.
引用
收藏
页码:273 / 283
页数:11
相关论文
共 50 条
  • [21] Facial Representation Extraction by Mutual Information Maximization and Correlation Minimization
    Wang, Xiaobo
    Sun, Wenyun
    Jin, Zhong
    INTELLIGENT INFORMATION AND DATABASE SYSTEMS, ACIIDS 2021, 2021, 12672 : 811 - 823
  • [22] Simple and Stable Internal Representation by Potential Mutual Information Maximization
    Kamimura, Ryotaro
    ENGINEERING APPLICATIONS OF NEURAL NETWORKS, EANN 2016, 2016, 629 : 309 - 316
  • [23] Unsupervised Visual Representation Learning via Mutual Information Regularized Assignment
    Lee, Dong Hoon
    Choi, Sungik
    Kim, Hyunwoo
    Chung, Sae-Young
    ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 35, NEURIPS 2022, 2022,
  • [24] Granular intents learning via mutual information maximization for knowledge-aware recommendation
    Yang, Hyeongjun
    Lee, Yerim
    Park, Gayeon
    Kim, Taeyoung
    Kim, Heesun
    Lee, Kyong-Ho
    Oh, Byungkook
    KNOWLEDGE-BASED SYSTEMS, 2024, 306
  • [25] Community detection in hypergraphs via mutual information maximization
    Kritschgau, Juergen
    Kaiser, Daniel
    Rodriguez, Oliver Alvarado
    Amburg, Ilya
    Bolkema, Jessalyn
    Grubb, Thomas
    Lan, Fangfei
    Maleki, Sepideh
    Chodrow, Phil
    Kay, Bill
    SCIENTIFIC REPORTS, 2024, 14 (01)
  • [26] Loop Closure Detection via Maximization of Mutual Information
    Zhang, Ge
    Yan, Xiaoqiang
    Ye, Yangdong
    IEEE ACCESS, 2019, 7 : 124217 - 124232
  • [27] Bipartite Graph Embedding via Mutual Information Maximization
    Cao, Jiangxia
    Lin, Xixun
    Guo, Shu
    Liu, Luchen
    Liu, Tingwen
    Wang, Bin
    WSDM '21: PROCEEDINGS OF THE 14TH ACM INTERNATIONAL CONFERENCE ON WEB SEARCH AND DATA MINING, 2021, : 635 - 643
  • [28] Online Continual Learning through Mutual Information Maximization
    Guo, Yidou
    Liu, Bing
    Zhao, Dongyan
    INTERNATIONAL CONFERENCE ON MACHINE LEARNING, VOL 162, 2022,
  • [29] Mutual Information Analysis in Multimodal Learning Systems
    Hadizadeh, Hadi
    Yeganli, S. Faegheh
    Rashidi, Bahador
    Bajic, Ivan V.
    2024 IEEE 7TH INTERNATIONAL CONFERENCE ON MULTIMEDIA INFORMATION PROCESSING AND RETRIEVAL, MIPR 2024, 2024, : 390 - 395
  • [30] Multiagent Reinforcement Learning With Graphical Mutual Information Maximization
    Ding, Shifei
    Du, Wei
    Ding, Ling
    Zhang, Jian
    Guo, Lili
    An, Bo
    IEEE TRANSACTIONS ON NEURAL NETWORKS AND LEARNING SYSTEMS, 2023, : 1 - 10