Scaling and Adapting Large Language Models for Portuguese Open Information Extraction: A Comparative Study of Fine-Tuning and LoRA

被引:0
|
作者
Melo, Alan [1 ]
Cabral, Bruno [1 ]
Claro, Daniela Barreiro [1 ]
机构
[1] Univ Fed Bahia, FORMAS Res Ctr Data & Nat Language, Inst Comp, Salvador, BA, Brazil
来源
关键词
OpenIE; Language Model; Information Extraction;
D O I
10.1007/978-3-031-79035-5_30
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
This paper comprehensively investigates the efficacy of different adaptation techniques for Large Language Models (LLMs) in the context of Open Information Extraction (OpenIE) for Portuguese. We compare Full Fine-Tuning (FFT) and Low-Rank Adaptation (LoRA) across a model with 0.5B parameters. Our study evaluates the impact of model size and adaptation method on OpenIE performance, considering precision, recall, and F1 scores, as well as computational efficiency during training and inference phases. We contribute to a high-performing LLM and novel insights into the trade-offs between model scale, adaptation technique, and cross-lingual transferability in the OpenIE task. Our findings reveal significant performance variations across different configurations, with LoRA demonstrating competitive results. We also analyze the linguistic nuances in the Portuguese OpenIE that pose challenges for models primarily trained on English data. This research advances our understanding of LLM adaptation for specialized NLP tasks and provides practical guidelines for deploying these models in resource-constrained and multilingual scenarios. Our work has implications for the broader cross-lingual open information extraction field and contributes to the ongoing discourse on efficient fine-tuning strategies for large pre-trained models.
引用
收藏
页码:427 / 441
页数:15
相关论文
共 50 条
  • [41] Characterizing Communication in Distributed Parameter-Efficient Fine-Tuning for Large Language Models
    Alnaasan, Nawras
    Huang, Horng-Ruey
    Shafi, Aamir
    Subramoni, Hari
    Panda, Dhabaleswar K.
    2024 IEEE SYMPOSIUM ON HIGH-PERFORMANCE INTERCONNECTS, HOTI 2024, 2024, : 11 - 19
  • [42] Data Selection for Fine-tuning Large Language Models Using Transferred Shapley Values
    Schoch, Stephanie
    Mishra, Ritwick
    Ji, Yangfeng
    PROCEEDINGS OF THE 61ST ANNUAL MEETING OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS, ACL-SRW 2023, VOL 4, 2023, : 266 - 275
  • [43] SPDF: Sparse Pre-training and Dense Fine-tuning for Large Language Models
    Thangarasa, Vithursan
    Gupta, Abhay
    Marshall, William
    Li, Tianda
    Leong, Kevin
    DeCoste, Dennis
    Lie, Sean
    Saxena, Shreyas
    UNCERTAINTY IN ARTIFICIAL INTELLIGENCE, 2023, 216 : 2134 - 2146
  • [44] Research on Fine-Tuning Optimization Strategies for Large Language Models in Tabular Data Processing
    Zhao, Xiaoyong
    Leng, Xingxin
    Wang, Lei
    Wang, Ningning
    BIOMIMETICS, 2024, 9 (11)
  • [45] Fine-tuning Pre-Trained Transformer Language Models to Distantly Supervised Relation Extraction
    Alt, Christoph
    Huebner, Marc
    Hennig, Leonhard
    57TH ANNUAL MEETING OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS (ACL 2019), 2019, : 1388 - 1398
  • [46] Empirical study on fine-tuning pre-trained large language models for fault diagnosis of complex systems
    Zheng, Shuwen
    Pan, Kai
    Liu, Jie
    Chen, Yunxia
    RELIABILITY ENGINEERING & SYSTEM SAFETY, 2024, 252
  • [47] Sample Size Considerations for Fine-Tuning Large Language Models for Named Entity Recognition Tasks: Methodological Study
    Majdik, Zoltan P.
    Graham, S. Scott
    Edward, Jade C. Shiva
    Rodriguez, Sabrina N.
    Karnes, Martha S.
    Jensen, Jared T.
    Barbour, Joshua B.
    Rousseau, Justin F.
    JMIR AI, 2024, 3
  • [48] A comparative study of fine-tuning deep learning models for plant disease identification
    Too, Edna Chebet
    Li Yujian
    Njuki, Sam
    Liu Yingchun
    COMPUTERS AND ELECTRONICS IN AGRICULTURE, 2019, 161 : 272 - 279
  • [49] Two-Stage Fine-Tuning for Improved Bias and Variance for Large Pretrained Language Models
    Wang, Lijing
    Li, Yingya
    Miller, Timothy
    Bethard, Steven
    Savova, Guergana
    PROCEEDINGS OF THE 61ST ANNUAL MEETING OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS (ACL 2023): LONG PAPERS, VOL 1, 2023, : 15746 - 15761
  • [50] Efficient Fine-Tuning of Large Language Models via a Low-Rank Gradient Estimator
    Zhang, Luoming
    Lou, Zhenyu
    Ying, Yangwei
    Yang, Cheng
    Zhou, Hong
    APPLIED SCIENCES-BASEL, 2025, 15 (01):