Explainable Natural Language Inference in the Legal Domain via Text Generation

被引:0
|
作者
Choi J. [1 ]
Honda U. [1 ]
Watanabe T. [1 ]
Inui K. [1 ]
机构
[1] Nara Institute of Science and Technology/RIKEN, Japan
关键词
legal; natural language inference; textual entailment recognition;
D O I
10.1527/tjsai.38-3_C-MB6
中图分类号
学科分类号
摘要
Natural language inference (NLI) in the legal domain is the task of predicting entailment between the premise, i.e. law, and the hypothesis, which is a statement regarding a legal issue. Current state-of-the-art approaches to NLI with pre-trained language models do not perform well in the legal domain, presumably due to a discrepancy in the level of abstraction between the premise and hypothesis and the convoluted nature of legal language. Some of the difficulties specific to the legal domain are that 1) the premise and hypothesis tend to be extensive in length; 2) the premise comprises multiple rules, and only one of the rules is related to the hypothesis. Thus only small fractions of the statements are relevant for determining entailment, while the rest is noise, and; 3) the premise is often abstract and written in legal terms, whereas the hypothesis is a concrete case and tends to be written with more ordinary vocabulary. These problems are accentuated by the scarcity of such data in the legal domain due to the high cost. Pretrained language models have been shown to be effective on natural language inference tasks in the legal domain. However, previous methods do not provide an explanation for the decisions, which is especially desirable in knowledge-intensive domains such as law. This study proposes to leverage the characteristics of legal texts and decomposes the overall NLI task into two simpler sub-steps. Specifically, we regard the hypothesis as a pair of a condition and a consequence and train a conditional language model to generate the consequence from a given premise and a condition. The trained model can be regarded as a knowledge source for generating a consequence given the query consisting of the premise and the condition. After that, when the model receives an entailment example, it should generate a consequence similar to the original consequence, and when it is a contradiction example, they should be dissimilar since the model is trained on entailment examples only. Then, we compare the generated consequence and the consequence part of the hypothesis to see whether they are similar or dissimilar by training a classifier. Experimental results on datasets derived from the Japanese bar exam show significant improvement in accuracy from prior methods. © 2023, Japanese Society for Artificial Intelligence. All rights reserved.
引用
收藏
相关论文
共 50 条
  • [21] Few-Shot Text Generation with Natural Language Instructions
    Schick, Timo
    Schuetze, Hinrich
    2021 CONFERENCE ON EMPIRICAL METHODS IN NATURAL LANGUAGE PROCESSING (EMNLP 2021), 2021, : 390 - 402
  • [22] Generation of natural language text using perspective descriptor in frames
    Uma, GV
    Geetha, TV
    IETE JOURNAL OF RESEARCH, 2001, 47 (1-2) : 43 - 57
  • [23] ALTER: Auxiliary Text Rewriting Tool for Natural Language Generation
    Xu, Qiongkai
    Xu, Chenchen
    Qu, Lizhen
    2019 CONFERENCE ON EMPIRICAL METHODS IN NATURAL LANGUAGE PROCESSING AND THE 9TH INTERNATIONAL JOINT CONFERENCE ON NATURAL LANGUAGE PROCESSING (EMNLP-IJCNLP 2019): PROCEEDINGS OF SYSTEM DEMONSTRATIONS, 2019, : 13 - 18
  • [24] Exploiting Cloze Questions for Few Shot Text Classification and Natural Language Inference
    Schick, Timo
    Schutze, Hinrich
    16TH CONFERENCE OF THE EUROPEAN CHAPTER OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS (EACL 2021), 2021, : 255 - 269
  • [25] Natural language inference for curation of structured clinical registries from unstructured text
    Percha, Bethany
    Pisapati, Kereeti
    Gao, Cynthia
    Schmidt, Hank
    JOURNAL OF THE AMERICAN MEDICAL INFORMATICS ASSOCIATION, 2021, 29 (01) : 97 - 108
  • [26] Explainable Text Classification in Legal Document Review A Case Study of Explainable Predictive Coding
    Chhatwal, Rishi
    Gronvall, Peter
    Huber-Fliflet, Nathaniel
    Keeling, Robert
    Zhang, Jianping
    Zhao, Haozhen
    2018 IEEE INTERNATIONAL CONFERENCE ON BIG DATA (BIG DATA), 2018, : 1905 - 1911
  • [27] Mining and Indexing of Legal Natural Language Texts with Domain and Task Ontology
    Kurcheeva, Galina
    Rakhvalova, Marina
    Rakhvalova, Daria
    Bakaev, Maxim
    ELECTRONIC GOVERNANCE AND OPEN SOCIETY: CHALLENGES IN EURASIA, EGOSE 2018, 2019, 947 : 123 - 137
  • [28] Unifying Vision-and-Language Tasks via Text Generation
    Cho, Jaemin
    Lei, Jie
    Tan, Hao
    Bansal, Mohit
    INTERNATIONAL CONFERENCE ON MACHINE LEARNING, VOL 139, 2021, 139
  • [29] Hybrid explainable image caption generation using image processing and natural language processing
    Mishra, Atul
    Agrawal, Anubhav
    Bhasker, Shailendra
    INTERNATIONAL JOURNAL OF SYSTEM ASSURANCE ENGINEERING AND MANAGEMENT, 2024, 15 (10) : 4874 - 4884
  • [30] Text visualization for geological hazard documents via text mining and natural language processing
    Ying Ma
    Zhong Xie
    Gang Li
    Kai Ma
    Zhen Huang
    Qinjun Qiu
    Hui Liu
    Earth Science Informatics, 2022, 15 : 439 - 454