On Effectiveness of Further Pre-training on BERT Models for Story Point Estimation

被引:0
|
作者
Amasaki, Sousuke [1 ]
机构
[1] Okayama Prefectural Univ, Dept Syst Engn, Soja, Okayama, Japan
关键词
effort estimation; BERT; further pre-training; story points;
D O I
10.1145/3617555.3617877
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
CONTEXT: Recent studies on story point estimation used deep learning-based language models. These language models were pre-trained on general corpora. However, using language models further pre-trained with specific corpora might be effective. OBJECTIVE: To examine the effectiveness of further pre-trained language models for the predictive performance of story point estimation. METHOD: Two types of further pre-trained language models, namely, domain-specific and repository-specific models, were compared with off-the-shelf models and Deep-SE. The estimation performance was evaluated with 16 project data. RESULTS: The effectiveness of domain-specific and repository-specific models were limited though they were better than the base model they further pre-trained. CONCLUSION: The effect of further pre-training was small. Large off-the-shelf models might be better to be chosen.
引用
收藏
页码:49 / 53
页数:5
相关论文
共 50 条
  • [21] Dom-BERT: Detecting Malicious Domains with Pre-training Model
    Tian, Yu
    Li, Zhenyu
    PASSIVE AND ACTIVE MEASUREMENT, PAM 2024, PT I, 2024, 14537 : 133 - 158
  • [22] Dict-BERT: Enhancing Language Model Pre-training with Dictionary
    Yu, Wenhao
    Zhu, Chenguang
    Fang, Yuwei
    Yu, Donghan
    Wang, Shuohang
    Xu, Yichong
    Zeng, Michael
    Jiang, Meng
    FINDINGS OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS (ACL 2022), 2022, : 1907 - 1918
  • [23] U-BERT: Pre-training User Representations for Improved Recommendation
    Qiu, Zhaopeng
    Wu, Xian
    Gao, Jingyue
    Fan, Wei
    THIRTY-FIFTH AAAI CONFERENCE ON ARTIFICIAL INTELLIGENCE, THIRTY-THIRD CONFERENCE ON INNOVATIVE APPLICATIONS OF ARTIFICIAL INTELLIGENCE AND THE ELEVENTH SYMPOSIUM ON EDUCATIONAL ADVANCES IN ARTIFICIAL INTELLIGENCE, 2021, 35 : 4320 - 4327
  • [24] Affect Analysis in Arabic Text: Further Pre-Training Language Models for Sentiment and Emotion
    Alshehri, Wafa
    Al-Twairesh, Nora
    Alothaim, Abdulrahman
    APPLIED SCIENCES-BASEL, 2023, 13 (09):
  • [25] EBFP: Example-Based Further Pre-training
    Zhang, Yanan
    Wu, Chaofan
    Zhao, Wang
    Zhang, Xiankun
    ADVANCED INTELLIGENT COMPUTING TECHNOLOGY AND APPLICATIONS, PT III, ICIC 2024, 2024, 14877 : 427 - 437
  • [26] Pre-training Mention Representations in Coreference Models
    Varkel, Yuval
    Globerson, Amir
    PROCEEDINGS OF THE 2020 CONFERENCE ON EMPIRICAL METHODS IN NATURAL LANGUAGE PROCESSING (EMNLP), 2020, : 8534 - 8540
  • [27] Pre-training Language Models for Comparative Reasoning
    Yu, Mengxia
    Zhang, Zhihan
    Yu, Wenhao
    Jiang, Meng
    2023 CONFERENCE ON EMPIRICAL METHODS IN NATURAL LANGUAGE PROCESSING (EMNLP 2023), 2023, : 12421 - 12433
  • [28] VILA: On Pre-training for Visual Language Models
    Lin, Ji
    Yin, Hongxu
    Ping, Wei
    Molchanov, Pavlo
    Shoeybi, Mohammad
    Han, Song
    2024 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR), 2024, : 26679 - 26689
  • [29] Kaleido-BERT: Vision-Language Pre-training on Fashion Domain
    Zhuge, Mingchen
    Gao, Dehong
    Fan, Deng-Ping
    Jin, Linbo
    Chen, Ben
    Zhou, Haoming
    Qiu, Minghui
    Shao, Ling
    2021 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION, CVPR 2021, 2021, : 12642 - 12652
  • [30] Geo-BERT Pre-training Model for Query Rewriting in POI Search
    Liu, Xiao
    Hu, Juan
    Shen, Qi
    Chen, Huan
    FINDINGS OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS, EMNLP 2021, 2021, : 2209 - 2214