Using Bidirectional Encoder Representations from Transformers (BERT) to predict criminal charges and sentences from Taiwanese court judgments

被引:1
|
作者
Peng, Yi-Ting [1 ]
Lei, Chin-Laung [1 ]
机构
[1] Natl Taiwan Univ, Dept Elect Engn, Taipei City, Taiwan
关键词
National language processing; Bidirectional encoder representations from transformer (BERT); Legal artificial intelligence (Legal AI); Artificial intelligence law (AI Law); Legal judgment prediction;
D O I
10.7717/peerj-cs.1841
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
People unfamiliar with the law may not know what kind of behavior is considered criminal behavior or the lengths of sentences tied to those behaviors. This study used criminal judgments from the district court in Taiwan to predict the type of crime and sentence length that would be determined. This study pioneers using Taiwanese criminal judgments as a dataset and proposes improvements based on Bidirectional Encoder Representations from Transformers (BERT). This study is divided into two parts: criminal charges prediction and sentence prediction. Injury and public endangerment judgments were used as training data to predict sentences. This study also proposes an effective solution to BERT's 512 -token limit. The results show that using the BERT model to train Taiwanese criminal judgments is feasible. Accuracy reached 98.95% in predicting criminal charges and 72.37% in predicting the sentence in injury trials, and 80.93% in predicting the sentence in public endangerment trials.
引用
收藏
页数:24
相关论文
共 50 条
  • [41] Predicting Antimalarial Activity in Natural Products Using Pretrained Bidirectional Encoder Representations from Transformers
    Nguyen-Vo, Thanh-Hoang
    Trinh, Quang H.
    Nguyen, Loc
    Do, Trang T. T.
    Chua, Matthew Chin Heng
    Nguyen, Binh P.
    JOURNAL OF CHEMICAL INFORMATION AND MODELING, 2022, 62 (21) : 5050 - 5058
  • [42] Adapting Bidirectional Encoder Representations from Transformers (BERT) to Assess Clinical Semantic Textual Similarity: Algorithm Development and Validation Study
    Kades, Klaus
    Sellner, Jan
    Koehler, Gregor
    Full, Peter M.
    Lai, T. Y. Emmy
    Kleesiek, Jens
    Maier-Hein, Klaus H.
    JMIR MEDICAL INFORMATICS, 2021, 9 (02)
  • [43] Sentiment Analysis of Turkish Drug Reviews with Bidirectional Encoder Representations from Transformers
    Bozuyla, Mehmet
    ACM TRANSACTIONS ON ASIAN AND LOW-RESOURCE LANGUAGE INFORMATION PROCESSING, 2024, 23 (01)
  • [44] ADVANCED TURKISH FAKE NEWS PREDICTION WITH BIDIRECTIONAL ENCODER REPRESENTATIONS FROM TRANSFORMERS
    Bozuyla, Mehmet
    KONYA JOURNAL OF ENGINEERING SCIENCES, 2022, 10 (03): : 750 - 761
  • [45] ECC-BERT: Classification of error correcting codes using the improved bidirectional encoder representation from transformers
    Li, Sida
    Hu, Xiaochang
    Huang, Zhiping
    Zhou, Jing
    IET COMMUNICATIONS, 2022, 16 (04) : 359 - 368
  • [46] BERT4Rec: Sequential Recommendation with Bidirectional Encoder Representations from Transformer
    Sun, Fei
    Liu, Jun
    Wu, Jian
    Pei, Changhua
    Lin, Xiao
    Ou, Wenwu
    Jiang, Peng
    PROCEEDINGS OF THE 28TH ACM INTERNATIONAL CONFERENCE ON INFORMATION & KNOWLEDGE MANAGEMENT (CIKM '19), 2019, : 1441 - 1450
  • [47] Bilingual Question Answering System Using Bidirectional Encoder Representations from Transformers and Best Matching Method
    Navastara, Dini Adni
    Ihdiannaja
    Arifm, Agus Zainal
    PROCEEDINGS OF 2021 13TH INTERNATIONAL CONFERENCE ON INFORMATION & COMMUNICATION TECHNOLOGY AND SYSTEM (ICTS), 2021, : 360 - 364
  • [48] Prediction of Machine-Generated Financial Tweets Using Advanced Bidirectional Encoder Representations from Transformers
    Arshed, Muhammad Asad
    Gherghina, Stefan Cristian
    Dur-E-Zahra
    Manzoor, Mahnoor
    ELECTRONICS, 2024, 13 (11)
  • [49] Using Data Augmentation and Bidirectional Encoder Representations from Transformers for Improving Punjabi Named Entity Recognition
    Khalid, Hamza
    Murtaza, Ghulam
    Abbas, Qaiser
    ACM TRANSACTIONS ON ASIAN AND LOW-RESOURCE LANGUAGE INFORMATION PROCESSING, 2023, 22 (06)
  • [50] Contextual classification of clinical records with bidirectional long short-term memory (Bi-LSTM) and bidirectional encoder representations from transformers (BERT) model
    Zalte, Jaya
    Shah, Harshal
    COMPUTATIONAL INTELLIGENCE, 2024, 40 (04)