Finding and Editing Multi-Modal Neurons in Pre-Trained Transformers

被引:0
|
作者
Pan, Haowen [1 ]
Cao, Yixin [2 ]
Wang, Xiaozhi [3 ]
Yang, Xun [1 ]
Wang, Meng [4 ]
机构
[1] Univ Sci & Technol China, Hefei, Peoples R China
[2] Fudan Univ, Sch Comp Sci, Shanghai, Peoples R China
[3] Tsinghua Univ, Beijing, Peoples R China
[4] Hefei Univ Technol, Hefei, Peoples R China
基金
中国国家自然科学基金;
关键词
D O I
暂无
中图分类号
学科分类号
摘要
Understanding the internal mechanisms by which multi-modal large language models (LLMs) interpret different modalities and integrate cross-modal representations is becoming increasingly critical for continuous improvements in both academia and industry. In this paper, we propose a novel method to identify key neurons for interpretability- how multi-modal LLMs bridge visual and textual concepts for captioning. Our method improves conventional works upon efficiency and applied range by removing needs of costly gradient computation. Based on those identified neurons, we further design a multi-modal knowledge editing method, beneficial to mitigate sensitive words or hallucination. For rationale of our design, we provide theoretical assumption. For empirical evaluation, we have conducted extensive quantitative and qualitative experiments. The results not only validate the effectiveness of our methods, but also offer insightful findings that highlight three key properties of multi-modal neurons: sensitivity, specificity and causal-effect, to shed light for future research.(1)
引用
收藏
页码:1012 / 1037
页数:26
相关论文
共 50 条
  • [1] PMMN: Pre-trained multi-Modal network for scene text recognition
    Zhang, Yu
    Fu, Zilong
    Huang, Fuyu
    Liu, Yizhi
    PATTERN RECOGNITION LETTERS, 2021, 151 : 103 - 111
  • [2] Probing Multi-modal Machine Translation with Pre-trained Language Model
    Kong, Yawei
    Fan, Kai
    FINDINGS OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS, ACL-IJCNLP 2021, 2021, : 3689 - 3699
  • [3] Large-scale Multi-modal Pre-trained Models: A Comprehensive Survey
    Xiao Wang
    Guangyao Chen
    Guangwu Qian
    Pengcheng Gao
    Xiao-Yong Wei
    Yaowei Wang
    Yonghong Tian
    Wen Gao
    Machine Intelligence Research, 2023, 20 : 447 - 482
  • [4] Large-scale Multi-modal Pre-trained Models: A Comprehensive Survey
    Wang, Xiao
    Chen, Guangyao
    Qian, Guangwu
    Gao, Pengcheng
    Wei, Xiao-Yong
    Wang, Yaowei
    Tian, Yonghong
    Gao, Wen
    MACHINE INTELLIGENCE RESEARCH, 2023, 20 (04) : 447 - 482
  • [5] Are Pre-trained Convolutions Better than Pre-trained Transformers?
    Tay, Yi
    Dehghani, Mostafa
    Gupta, Jai
    Aribandi, Vamsi
    Bahri, Dara
    Qin, Zhen
    Metzler, Donald
    59TH ANNUAL MEETING OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS AND THE 11TH INTERNATIONAL JOINT CONFERENCE ON NATURAL LANGUAGE PROCESSING (ACL-IJCNLP 2021), VOL 1, 2021, : 4349 - 4359
  • [6] MultiFusion: Fusing Pre-Trained Models for Multi-Lingual, Multi-Modal Image Generation
    Bellagente, Marco
    Brack, Manuel
    Teufel, Hannah
    Friedrich, Felix
    Deiseroth, Bjoern
    Eichenberg, Constantin
    Dai, Andrew
    Baldock, Robert J. N.
    Nanda, Souradeep
    Oostermeijer, Koen
    Cruz-Salinas, Andres Felipe
    Schramowski, Patrick
    Kersting, Kristian
    Weinbach, Samuel
    ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 36 (NEURIPS 2023), 2023,
  • [7] Multi-modal Segmentation with Missing MR Sequences Using Pre-trained Fusion Networks
    van Garderen, Karin
    Smits, Marion
    Klein, Stefan
    DOMAIN ADAPTATION AND REPRESENTATION TRANSFER AND MEDICAL IMAGE LEARNING WITH LESS LABELS AND IMPERFECT DATA, DART 2019, MIL3ID 2019, 2019, 11795 : 165 - 172
  • [8] Calibration of Pre-trained Transformers
    Desai, Shrey
    Durrett, Greg
    PROCEEDINGS OF THE 2020 CONFERENCE ON EMPIRICAL METHODS IN NATURAL LANGUAGE PROCESSING (EMNLP), 2020, : 295 - 302
  • [9] Cyberbullying detection on multi-modal data using pre-trained deep learning architectures
    Pericherla, Subbaraju
    Ilavarasan, E.
    INGENIERIA SOLIDARIA, 2021, 17 (03):
  • [10] Fast multi-modal reuse: Co-occurrence pre-trained deep learning models
    Iyer, Vasanth
    Aved, Alexander
    Howlett, Todd B.
    Carlo, Jeffrey T.
    Mehmood, Asif
    Pissinou, Niki
    Iyengar, S.S.
    Proceedings of SPIE - The International Society for Optical Engineering, 2019, 10996