Word Representation Learning in Multimodal Pre-Trained Transformers: An Intrinsic Evaluation

被引:4
|
作者
Pezzelle, Sandro [1 ]
Takmaz, Ece [1 ]
Fernandez, Raquel [1 ]
机构
[1] Univ Amsterdam, Inst Log Language & Computat, Amsterdam, Netherlands
基金
欧洲研究理事会;
关键词
DISTRIBUTIONAL SEMANTICS; MODELS;
D O I
10.1162/tacl_a_00443
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
This study carries out a systematic intrinsic evaluation of the semantic representations learned by state-of-the-art pre-trained multimodal Transformers. These representations are claimed to be task-agnostic and shown to help on many downstream language-and-vision tasks. However, the extent to which they align with human semantic intuitions remains unclear. We experiment with various models and obtain static word representations from the contextualized ones they learn. We then evaluate them against the semantic judgments provided by human speakers. In linewith previous evidence, we observe a generalized advantage of multimodal representations over languageonly ones on concrete word pairs, but not on abstract ones. On the one hand, this confirms the effectiveness of these models to align language and vision, which results in better semantic representations for concepts that are grounded in images. On the other hand, models are shown to follow different representation learning patterns, which sheds some light on how and when they perform multimodal integration.
引用
收藏
页码:1563 / 1579
页数:17
相关论文
共 50 条
  • [21] Predicting Terms in IS-A Relations with Pre-trained Transformers
    Nikishina, Irina
    Chernomorchenko, Polina
    Demidova, Anastasiia
    Panchenko, Alexander
    Biemann, Chris
    13TH INTERNATIONAL JOINT CONFERENCE ON NATURAL LANGUAGE PROCESSING AND THE 3RD CONFERENCE OF THE ASIA-PACIFIC CHAPTER OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS, IJCNLP-AACL 2023, 2023, : 134 - 148
  • [22] Efficient feature selection for pre-trained vision transformers
    Huang, Lan
    Zeng, Jia
    Yu, Mengqiang
    Ding, Weiping
    Bai, Xingyu
    Wang, Kangping
    COMPUTER VISION AND IMAGE UNDERSTANDING, 2025, 254
  • [23] Generative pre-trained transformers (GPT) for surface engineering
    Kamnis, Spyros
    SURFACE & COATINGS TECHNOLOGY, 2023, 466
  • [24] Generating Extended and Multilingual Summaries with Pre-trained Transformers
    Calizzano, Remi
    Ostendorff, Malte
    Ruan, Qian
    Rehm, Georg
    LREC 2022: THIRTEEN INTERNATIONAL CONFERENCE ON LANGUAGE RESOURCES AND EVALUATION, 2022, : 1640 - 1650
  • [25] GENERATIVE PRE-TRAINED TRANSFORMERS FOR BIOLOGICALLY INSPIRED DESIGN
    Zhu, Qihao
    Zhang, Xinyu
    Luo, Jianxi
    PROCEEDINGS OF ASME 2022 INTERNATIONAL DESIGN ENGINEERING TECHNICAL CONFERENCES AND COMPUTERS AND INFORMATION IN ENGINEERING CONFERENCE, IDETC-CIE2022, VOL 6, 2022,
  • [26] A Robust Representation with Pre-trained Start and End Characters Vectors for Noisy Word Recognition
    Liu, Chao
    Ma, Xiangmei
    Yu, Min
    Wu, Xinghua
    Liu, Mingqi
    Jiang, Jianguo
    Huang, Weiqing
    KNOWLEDGE SCIENCE, ENGINEERING AND MANAGEMENT (KSEM 2020), PT I, 2020, 12274 : 174 - 185
  • [27] Fusing Pre-trained Language Models with Multimodal Prompts through Reinforcement Learning
    Yu, Youngjae
    Chung, Jiwan
    Yun, Heeseung
    Hessel, Jack
    Park, Jae Sung
    Lu, Ximing
    Zellers, Rowan
    Ammanabrolu, Prithviraj
    Le Bras, Ronan
    Kim, Gunhee
    Choi, Yejin
    2023 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR), 2023, : 10845 - 10856
  • [28] Learning Social Relationship From Videos via Pre-Trained Multimodal Transformer
    Teng, Yiyang
    Song, Chenguang
    Wu, Bin
    IEEE SIGNAL PROCESSING LETTERS, 2022, 29 : 1377 - 1381
  • [29] Ensemble Learning with Pre-Trained Transformers for Crash Severity Classification: A Deep NLP Approach
    Jaradat, Shadi
    Nayak, Richi
    Paz, Alexander
    Elhenawy, Mohammed
    ALGORITHMS, 2024, 17 (07)
  • [30] PART: Pre-trained Authorship Representation Transformer
    Huertas-Tato, Javier
    Martin, Alejandro
    Camacho, David
    HUMAN-CENTRIC COMPUTING AND INFORMATION SCIENCES, 2024, 14