Explainable Natural Language Inference in the Legal Domain via Text Generation

被引:0
|
作者
Choi J. [1 ]
Honda U. [1 ]
Watanabe T. [1 ]
Inui K. [1 ]
机构
[1] Nara Institute of Science and Technology/RIKEN, Japan
关键词
legal; natural language inference; textual entailment recognition;
D O I
10.1527/tjsai.38-3_C-MB6
中图分类号
学科分类号
摘要
Natural language inference (NLI) in the legal domain is the task of predicting entailment between the premise, i.e. law, and the hypothesis, which is a statement regarding a legal issue. Current state-of-the-art approaches to NLI with pre-trained language models do not perform well in the legal domain, presumably due to a discrepancy in the level of abstraction between the premise and hypothesis and the convoluted nature of legal language. Some of the difficulties specific to the legal domain are that 1) the premise and hypothesis tend to be extensive in length; 2) the premise comprises multiple rules, and only one of the rules is related to the hypothesis. Thus only small fractions of the statements are relevant for determining entailment, while the rest is noise, and; 3) the premise is often abstract and written in legal terms, whereas the hypothesis is a concrete case and tends to be written with more ordinary vocabulary. These problems are accentuated by the scarcity of such data in the legal domain due to the high cost. Pretrained language models have been shown to be effective on natural language inference tasks in the legal domain. However, previous methods do not provide an explanation for the decisions, which is especially desirable in knowledge-intensive domains such as law. This study proposes to leverage the characteristics of legal texts and decomposes the overall NLI task into two simpler sub-steps. Specifically, we regard the hypothesis as a pair of a condition and a consequence and train a conditional language model to generate the consequence from a given premise and a condition. The trained model can be regarded as a knowledge source for generating a consequence given the query consisting of the premise and the condition. After that, when the model receives an entailment example, it should generate a consequence similar to the original consequence, and when it is a contradiction example, they should be dissimilar since the model is trained on entailment examples only. Then, we compare the generated consequence and the consequence part of the hypothesis to see whether they are similar or dissimilar by training a classifier. Experimental results on datasets derived from the Japanese bar exam show significant improvement in accuracy from prior methods. © 2023, Japanese Society for Artificial Intelligence. All rights reserved.
引用
收藏
相关论文
共 50 条
  • [31] Text visualization for geological hazard documents via text mining and natural language processing
    Ma, Ying
    Xie, Zhong
    Li, Gang
    Ma, Kai
    Huang, Zhen
    Qiu, Qinjun
    Liu, Hui
    EARTH SCIENCE INFORMATICS, 2022, 15 (01) : 439 - 454
  • [32] Natural prosody generation for domain specific text-to-speech systems
    Katae, N
    Kimura, S
    ICSLP 96 - FOURTH INTERNATIONAL CONFERENCE ON SPOKEN LANGUAGE PROCESSING, PROCEEDINGS, VOLS 1-4, 1996, : 1852 - 1855
  • [33] Genetic algorithm based sentence packaging in natural language text generation
    Devyatkin, Dmitry
    Isakov, Vadim
    Shvets, Alexander
    INTERNATIONAL WORKSHOP ADVANCED TECHNOLOGIES IN MATERIAL SCIENCE, MECHANICAL AND AUTOMATION ENGINEERING - MIP: ENGINEERING - 2019, 2019, 537
  • [34] Temporal Natural Language Inference: Evidence-Based Evaluation of Temporal Text Validity
    Hosokawa, Taishi
    Jatowt, Adam
    Sugiyama, Kazunari
    ADVANCES IN INFORMATION RETRIEVAL, ECIR 2023, PT I, 2023, 13980 : 441 - 458
  • [35] Causal Knowledge Extraction from Text using Natural Language Inference (Student Abstract)
    Bhandari, Manik
    Feblowitz, Mark
    Hassanzadeh, Oktie
    Srinivas, Kavitha
    Sohrabi, Shirin
    THIRTY-FIFTH AAAI CONFERENCE ON ARTIFICIAL INTELLIGENCE, THIRTY-THIRD CONFERENCE ON INNOVATIVE APPLICATIONS OF ARTIFICIAL INTELLIGENCE AND THE ELEVENTH SYMPOSIUM ON EDUCATIONAL ADVANCES IN ARTIFICIAL INTELLIGENCE, 2021, 35 : 15759 - 15760
  • [36] Improving Natural Language Inference Using External Knowledge in the Science Questions Domain
    Wang, Xiaoyan
    Kapanipathi, Pavan
    Musa, Ryan
    Yu, Mo
    Talamadupula, Kartik
    Abdelaziz, Ibrahim
    Chang, Maria
    Fokoue, Achille
    Makni, Bassem
    Mattei, Nicholas
    Witbrock, Michael
    THIRTY-THIRD AAAI CONFERENCE ON ARTIFICIAL INTELLIGENCE / THIRTY-FIRST INNOVATIVE APPLICATIONS OF ARTIFICIAL INTELLIGENCE CONFERENCE / NINTH AAAI SYMPOSIUM ON EDUCATIONAL ADVANCES IN ARTIFICIAL INTELLIGENCE, 2019, : 7208 - 7215
  • [37] Improved Natural Language Generation via Loss Truncation
    Kang, Daniel
    Hashimoto, Tatsunori B.
    58TH ANNUAL MEETING OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS (ACL 2020), 2020, : 718 - 731
  • [38] Unsupervised Natural Language Inference via Decoupled Multimodal Contrastive Learning
    Cui, Wanyun
    Zheng, Guangyu
    Wang, Wei
    PROCEEDINGS OF THE 2020 CONFERENCE ON EMPIRICAL METHODS IN NATURAL LANGUAGE PROCESSING (EMNLP), 2020, : 5511 - 5520
  • [39] Arabic Text Classification in the Legal Domain
    Ait Yahia, Ikram
    Loqman, Chakir
    2019 THIRD INTERNATIONAL CONFERENCE ON INTELLIGENT COMPUTING IN DATA SCIENCES (ICDS 2019), 2019,
  • [40] Text Classification and Prediction in the Legal Domain
    Nghiem, Minh-Quoc
    Baylis, Paul
    Freitas, Andre
    Ananiadou, Sophia
    LREC 2022: THIRTEEN INTERNATIONAL CONFERENCE ON LANGUAGE RESOURCES AND EVALUATION, 2022, : 4717 - 4722