Local Interpretations for Explainable Natural Language Processing: A Survey

被引:7
|
作者
Luo, Siwen [1 ]
Ivison, Hamish [2 ]
Han, Soyeon Caren [3 ]
Poon, Josiah [4 ]
机构
[1] Univ Western Australia, 35 Stirling Hwy, Perth, WA 6009, Australia
[2] Univ Washington, 3800 E Stevens Way NE, Seattle, WA 98195 USA
[3] Univ Melbourne, 700 Swanston St, Melbourne, Vic 3010, Australia
[4] Univ Sydney, 1 Cleveland St, Darlington, NSW 2008, Australia
关键词
Deep neural networks; explainable AI; local interpretation; natural language processing; PREDICTION;
D O I
10.1145/3649450
中图分类号
TP301 [理论、方法];
学科分类号
081202 ;
摘要
As the use of deep learning techniques has grown across various fields over the past decade, complaints about the opaqueness of the black-box models have increased, resulting in an increased focus on transparency in deep learning models. This work investigates various methods to improve the interpretability of deep neural networks for Natural Language Processing (NLP) tasks, including machine translation and sentiment analysis. We provide a comprehensive discussion on the definition of the term interpretability and its various aspects at the beginning of this work. The methods collected and summarised in this survey are only associated with local interpretation and are specifically divided into three categories: (1) interpreting the model's predictions through related input features; (2) interpreting through natural language explanation; (3) probing the hidden states of models and word representations.
引用
收藏
页数:36
相关论文
共 50 条
  • [1] A Survey of the State of Explainable AI for Natural Language Processing
    Danilevsky, Marina
    Qian, Kun
    Aharonov, Ranit
    Katsis, Yannis
    Kawas, Ban
    Sen, Prithviraj
    1ST CONFERENCE OF THE ASIA-PACIFIC CHAPTER OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS AND THE 10TH INTERNATIONAL JOINT CONFERENCE ON NATURAL LANGUAGE PROCESSING (AACL-IJCNLP 2020), 2020, : 447 - 459
  • [2] Explainable Natural Language Processing
    Zhang, Zihao
    NATURAL LANGUAGE ENGINEERING, 2024, 30 (04) : 882 - 885
  • [3] Editorial: Explainable AI in Natural Language Processing
    Banerjee, Somnath
    Tomas, David
    FRONTIERS IN ARTIFICIAL INTELLIGENCE, 2024, 7
  • [4] Explainable natural language processing for corporate sustainability analysis
    Ong, Keane
    Mao, Rui
    Satapathy, Ranjan
    Shirota Filho, Ricardo
    Cambria, Erik
    Sulaeman, Johan
    Mengaldo, Gianmarco
    INFORMATION FUSION, 2025, 115
  • [5] Explainable natural language processing with matrix product states
    Tangpanitanon, Jirawat
    Mangkang, Chanatip
    Bhadola, Pradeep
    Minato, Yuichiro
    Angelakis, Dimitris G.
    Chotibut, Thiparat
    NEW JOURNAL OF PHYSICS, 2022, 24 (05):
  • [6] Natural Language Processing for Dialects of a Language: A Survey
    Joshi, Aditya
    Dabre, Raj
    Kanojia, Diptesh
    Li, Zhuang
    Zhan, Haolan
    Haffari, Gholamreza
    Dippold, Doris
    ACM COMPUTING SURVEYS, 2025, 57 (06)
  • [7] A survey of graphs in natural language processing
    Nastase, Vivi
    Mihalcea, Rada
    Radev, Dragomir R.
    NATURAL LANGUAGE ENGINEERING, 2015, 21 (05) : 665 - 698
  • [8] Natural language processing in finance: A survey
    Du, Kelvin
    Zhao, Yazhi
    Mao, Rui
    Xing, Frank
    Cambria, Erik
    INFORMATION FUSION, 2025, 115
  • [9] Hybrid explainable image caption generation using image processing and natural language processing
    Mishra, Atul
    Agrawal, Anubhav
    Bhasker, Shailendra
    INTERNATIONAL JOURNAL OF SYSTEM ASSURANCE ENGINEERING AND MANAGEMENT, 2024, 15 (10) : 4874 - 4884
  • [10] Quantum Natural Language Processing: A Comprehensive Survey
    Varmantchaonala, Charles M.
    Fendji, Jean Louis K. E.
    Schoning, Julius
    Atemkeng, Marcellin
    IEEE ACCESS, 2024, 12 : 99578 - 99598