Neural Data Augmentation for Legal Overruling Task: Small Deep Learning Models vs. Large Language Models

被引:0
|
作者
Sheik, Reshma [1 ]
Sundara, K. P. Siva [2 ]
Nirmala, S. Jaya [1 ]
机构
[1] Natl Inst Technol, Dept Comp Sci & Engn, Tiruchirapalli 620015, India
[2] Coimbatore Inst Technol, Dept Elect & Commun Engn, Coimbatore 641013, India
关键词
Deep learning; Natural language processing; Data augmentation; Legal overruling task; Transformer; Few-shot; GPT-3; Large language models;
D O I
10.1007/s11063-024-11574-4
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Deep learning models produce impressive results in any natural language processing applications when given a better learning strategy and trained with large labeled datasets. However, the annotation of massive training data is far too expensive, especially in the legal domain, due to the need for trained legal professionals. Data augmentation solves the problem of learning without labeled big data. In this paper, we employ pre-trained language models and prompt engineering to generate large-scale pseudo-labeled data for the legal overruling task using 100 data samples. We train small recurrent and convolutional deep-learning models using this data and fine-tune a few other transformer models. We then evaluate the effectiveness of the models, both with and without data augmentation, using the benchmark dataset and analyze the results. We also test the performance of these models with the state-of-the-art GPT-3 model under few-shot setting. Our experimental findings demonstrate that data augmentation results in better model performance in the legal overruling task than models trained without augmentation. Furthermore, our best-performing deep learning model trained on augmented data outperforms the few-shot GPT-3 by 18% in the F1-score. Additionally, our results highlight that the small neural networks trained with augmented data achieve outcomes comparable to those of other large language models.
引用
下载
收藏
页数:21
相关论文
共 50 条
  • [31] Shallow vs. Deep Learning Models for Groundwater Level Prediction: A Multi-Piezometer Data Integration Approach
    Yeganeh, Ali
    Ahmadi, Farshad
    Wong, Yong Jie
    Shadman, Alireza
    Barati, Reza
    Saeedi, Reza
    WATER AIR AND SOIL POLLUTION, 2024, 235 (07):
  • [32] Learning Higher Representations from pre-trained Deep Models with Data Augmentation for the COMPARE 2020 Challenge Mask Task
    Koike, Tomoya
    Qian, Kun
    Schuller, Bjoern W.
    Yamamoto, Yoshiharu
    INTERSPEECH 2020, 2020, : 2047 - 2051
  • [33] Advanced large language models and visualization tools for data analytics learning
    Valverde-Rebaza, Jorge
    Gonzalez, Aram
    Navarro-Hinojosa, Octavio
    Noguez, Julieta
    FRONTIERS IN EDUCATION, 2024, 9
  • [34] Integration of machine learning models for ADME to enable drug discovery: Deep neural network vs. support vector machine
    Desai, Prashant
    ABSTRACTS OF PAPERS OF THE AMERICAN CHEMICAL SOCIETY, 2019, 258
  • [35] Understanding Naturalistic Facial Expressions with Deep Learning and Multimodal Large Language Models
    Bian, Yifan
    Kuester, Dennis
    Liu, Hui
    Krumhuber, Eva G.
    SENSORS, 2024, 24 (01)
  • [36] A Proposed Approach for Object Detection and Recognition by Deep Learning Models Using Data Augmentation
    Abdulkareem, Ismael M.
    AL-Shammri, Faris K.
    Khalid, Noor Aldeen A.
    Omran, Natiq A.
    INTERNATIONAL JOURNAL OF ONLINE AND BIOMEDICAL ENGINEERING, 2024, 20 (05) : 31 - 43
  • [37] Effect of Data Augmentation Using Deep Learning on Predictive Models for Geopolymer Compressive Strength
    Nguyen, Ho Anh Thu
    Pham, Duy Hoang
    Ahn, Yonghan
    APPLIED SCIENCES-BASEL, 2024, 14 (09):
  • [38] Efficient deep learning models for brain tumor detection with segmentation and data augmentation techniques
    Shoaib, Mohamed R.
    Elshamy, Mohamed R.
    Taha, Taha E.
    El-Fishawy, Adel S.
    Abd El-Samie, Fathi E.
    CONCURRENCY AND COMPUTATION-PRACTICE & EXPERIENCE, 2022, 34 (21):
  • [39] RGB-Angle-Wheel: A new data augmentation method for deep learning models
    Ozdemir, Cuneyt
    Dogan, Yahya
    Kaya, Yilmaz
    KNOWLEDGE-BASED SYSTEMS, 2024, 291
  • [40] Open set task augmentation facilitates generalization of deep neural networks trained on small data sets
    Zai El Amri, Wadhah
    Reinhart, Felix
    Schenck, Wolfram
    NEURAL COMPUTING & APPLICATIONS, 2022, 34 (08): : 6067 - 6083