The unreasonable effectiveness of large language models in zero-shot semantic annotation of legal texts

被引:11
|
作者
Savelka, Jaromir [1 ]
Ashley, Kevin D. [2 ]
机构
[1] Carnegie Mellon Univ, Sch Comp Sci, Pittsburgh, PA 15213 USA
[2] Univ Pittsburgh, Sch Law, Pittsburgh, PA 15260 USA
来源
关键词
legal text analytics; large language models (LLM); zero-shot classification; semantic annotation; text annotation; CLASSIFICATION; EXTRACTION; DECISIONS; SEARCH;
D O I
10.3389/frai.2023.1279794
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
The emergence of ChatGPT has sensitized the general public, including the legal profession, to large language models' (LLMs) potential uses (e.g., document drafting, question answering, and summarization). Although recent studies have shown how well the technology performs in diverse semantic annotation tasks focused on legal texts, an influx of newer, more capable (GPT-4) or cost-effective (GPT-3.5-turbo) models requires another analysis. This paper addresses recent developments in the ability of LLMs to semantically annotate legal texts in zero-shot learning settings. Given the transition to mature generative AI systems, we examine the performance of GPT-4 and GPT-3.5-turbo(-16k), comparing it to the previous generation of GPT models, on three legal text annotation tasks involving diverse documents such as adjudicatory opinions, contractual clauses, or statutory provisions. We also compare the models' performance and cost to better understand the trade-offs. We found that the GPT-4 model clearly outperforms the GPT-3.5 models on two of the three tasks. The cost-effective GPT-3.5-turbo matches the performance of the 20x more expensive text-davinci-003 model. While one can annotate multiple data points within a single prompt, the performance degrades as the size of the batch increases. This work provides valuable information relevant for many practical applications (e.g., in contract review) and research projects (e.g., in empirical legal studies). Legal scholars and practicing lawyers alike can leverage these findings to guide their decisions in integrating LLMs in a wide range of workflows involving semantic annotation of legal texts.
引用
收藏
页数:14
相关论文
共 50 条
  • [41] EFFECTIVENESS OF ZERO-SHOT MODELS IN AUTOMATIC ARABIC POEM GENERATION
    Beheitt, Mohamed El Ghaly
    Ben HajHmida, Moez
    JORDANIAN JOURNAL OF COMPUTERS AND INFORMATION TECHNOLOGY, 2023, 9 (01): : 21 - 35
  • [42] Retrieving-to-Answer: Zero-Shot Video Question Answering with Frozen Large Language Models
    Pan, Junting
    Lin, Ziyi
    Ge, Yuying
    Zhu, Xiatian
    Zhang, Renrui
    Wang, Yi
    Qiao, Yu
    Li, Hongsheng
    2023 IEEE/CVF INTERNATIONAL CONFERENCE ON COMPUTER VISION WORKSHOPS, ICCVW, 2023, : 272 - 283
  • [43] CDZL: a controllable diversity zero-shot image caption model using large language models
    Zhao, Xin
    Kong, Weiwei
    Liu, Zongyao
    Wang, Menghao
    Li, Yiwen
    SIGNAL IMAGE AND VIDEO PROCESSING, 2025, 19 (04)
  • [44] Generating Training Data with Language Models: Towards Zero-Shot Language Understanding
    Meng, Yu
    Huang, Jiaxin
    Zhang, Yu
    Han, Jiawei
    ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 35, NEURIPS 2022, 2022,
  • [45] Zero-Shot Recommendation as Language Modeling
    Sileo, Damien
    Vossen, Wout
    Raymaekers, Robbe
    ADVANCES IN INFORMATION RETRIEVAL, PT II, 2022, 13186 : 223 - 230
  • [46] Towards Zero-shot Language Modeling
    Ponti, Edoardo M.
    Vulic, Ivan
    Cotterell, Ryan
    Reichart, Roi
    Korhonen, Anna
    2019 CONFERENCE ON EMPIRICAL METHODS IN NATURAL LANGUAGE PROCESSING AND THE 9TH INTERNATIONAL JOINT CONFERENCE ON NATURAL LANGUAGE PROCESSING (EMNLP-IJCNLP 2019): PROCEEDINGS OF THE CONFERENCE, 2019, : 2900 - +
  • [47] VLFM: Vision-Language Frontier Maps for Zero-Shot Semantic Navigation
    Yokoyama, Naoki
    Ha, Sehoon
    Batra, Dhruv
    Wang, Jiuguang
    Bucher, Bernadette
    2024 IEEE INTERNATIONAL CONFERENCE ON ROBOTICS AND AUTOMATION, ICRA 2024, 2024, : 42 - 48
  • [48] Open-source Large Language Models are Strong Zero-shot Query Likelihood Models for Document Ranking
    Zhuang, Shengyao
    Liu, Bing
    Koopman, Bevan
    Zuccon, Guido
    FINDINGS OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS (EMNLP 2023), 2023, : 8807 - 8817
  • [49] Learning semantic ambiguities for zero-shot learning
    Hanouti, Celina
    Le Borgne, Herve
    MULTIMEDIA TOOLS AND APPLICATIONS, 2023, 82 (26) : 40745 - 40759
  • [50] Learning semantic ambiguities for zero-shot learning
    Celina Hanouti
    Hervé Le Borgne
    Multimedia Tools and Applications, 2023, 82 : 40745 - 40759