Explainable Natural Language Inference in the Legal Domain via Text Generation

被引:0
|
作者
Choi J. [1 ]
Honda U. [1 ]
Watanabe T. [1 ]
Inui K. [1 ]
机构
[1] Nara Institute of Science and Technology/RIKEN, Japan
关键词
legal; natural language inference; textual entailment recognition;
D O I
10.1527/tjsai.38-3_C-MB6
中图分类号
学科分类号
摘要
Natural language inference (NLI) in the legal domain is the task of predicting entailment between the premise, i.e. law, and the hypothesis, which is a statement regarding a legal issue. Current state-of-the-art approaches to NLI with pre-trained language models do not perform well in the legal domain, presumably due to a discrepancy in the level of abstraction between the premise and hypothesis and the convoluted nature of legal language. Some of the difficulties specific to the legal domain are that 1) the premise and hypothesis tend to be extensive in length; 2) the premise comprises multiple rules, and only one of the rules is related to the hypothesis. Thus only small fractions of the statements are relevant for determining entailment, while the rest is noise, and; 3) the premise is often abstract and written in legal terms, whereas the hypothesis is a concrete case and tends to be written with more ordinary vocabulary. These problems are accentuated by the scarcity of such data in the legal domain due to the high cost. Pretrained language models have been shown to be effective on natural language inference tasks in the legal domain. However, previous methods do not provide an explanation for the decisions, which is especially desirable in knowledge-intensive domains such as law. This study proposes to leverage the characteristics of legal texts and decomposes the overall NLI task into two simpler sub-steps. Specifically, we regard the hypothesis as a pair of a condition and a consequence and train a conditional language model to generate the consequence from a given premise and a condition. The trained model can be regarded as a knowledge source for generating a consequence given the query consisting of the premise and the condition. After that, when the model receives an entailment example, it should generate a consequence similar to the original consequence, and when it is a contradiction example, they should be dissimilar since the model is trained on entailment examples only. Then, we compare the generated consequence and the consequence part of the hypothesis to see whether they are similar or dissimilar by training a classifier. Experimental results on datasets derived from the Japanese bar exam show significant improvement in accuracy from prior methods. © 2023, Japanese Society for Artificial Intelligence. All rights reserved.
引用
收藏
相关论文
共 50 条
  • [1] Explainable Natural Language Inference via Identifying Important Rationales
    Yang Z.
    Dong S.
    Hu J.
    IEEE Transactions on Artificial Intelligence, 2023, 4 (03): : 438 - 449
  • [2] Domain Specific Query Generation from Natural Language Text
    Iftikhar, Anum
    Iftikhar, Erum
    Mehmood, Muhammad Khalid
    2016 SIXTH INTERNATIONAL CONFERENCE ON INNOVATIVE COMPUTING TECHNOLOGY (INTECH), 2016, : 502 - 506
  • [3] Explaining Text Matching on Neural Natural Language Inference
    Kim, Youngwoo
    Jang, Myungha
    Allan, James
    ACM TRANSACTIONS ON INFORMATION SYSTEMS, 2020, 38 (04)
  • [4] SCINLI: A Corpus for Natural Language Inference on Scientific Text
    Sadat, Mobashir
    Caragea, Cornelia
    PROCEEDINGS OF THE 60TH ANNUAL MEETING OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS (ACL 2022), VOL 1: (LONG PAPERS), 2022, : 7399 - 7409
  • [5] On the Ethical Limits of Natural Language Processing on Legal Text
    Tsarapatsanis, Dimitrios
    Aletras, Nikolaos
    FINDINGS OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS, ACL-IJCNLP 2021, 2021, : 3590 - 3599
  • [6] Towards Explainable Search in Legal Text
    Polley, Sayantan
    ADVANCES IN INFORMATION RETRIEVAL, PT II, 2022, 13186 : 528 - 536
  • [7] Chinese Text Open Domain Tag Generation Method via Large Language Model
    He, Chunhui
    Ge, Bin
    Zhang, Chong
    2024 10TH INTERNATIONAL CONFERENCE ON BIG DATA AND INFORMATION ANALYTICS, BIGDIA 2024, 2024, : 183 - 188
  • [8] Lessons from Natural Language Inference in the Clinical Domain
    Romanov, Alexey
    Shivade, Chaitanya
    2018 CONFERENCE ON EMPIRICAL METHODS IN NATURAL LANGUAGE PROCESSING (EMNLP 2018), 2018, : 1586 - 1596
  • [9] Towards Cross-Domain Transferability of Text Generation Models for Legal Text
    Kumar, Vinayshekhar Bannihatti
    Bhattacharjee, Kasturi
    Gangadharaiah, Rashmi
    NLLP 2022 - Natural Legal Language Processing Workshop 2022, Proceedings of the Workshop, 2022, : 111 - 118
  • [10] A Review on Question Generation from Natural Language Text
    Zhang, Ruqing
    Guo, Jiafeng
    Chen, Lu
    Fan, Yixing
    Cheng, Xueqi
    ACM TRANSACTIONS ON INFORMATION SYSTEMS, 2022, 40 (01)