Automatic Taxonomy Classification by Pretrained Language Model

被引:0
|
作者
Kuwana, Ayato [1 ]
Oba, Atsushi [1 ]
Sawai, Ranto [1 ]
Paik, Incheon [1 ]
机构
[1] Univ Aizu, Grad Dept Comp Sci & Informat Syst, Fukushima, Fukui 9658580, Japan
关键词
ontology; automation; natural language processing (NLP); pretrained model;
D O I
10.3390/electronics10212656
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
In recent years, automatic ontology generation has received significant attention in information science as a means of systemizing vast amounts of online data. As our initial attempt of ontology generation with a neural network, we proposed a recurrent neural network-based method. However, updating the architecture is possible because of the development in natural language processing (NLP). By contrast, the transfer learning of language models trained by a large, unlabeled corpus has yielded a breakthrough in NLP. Inspired by these achievements, we propose a novel workflow for ontology generation comprising two-stage learning. Our results showed that our best method improved accuracy by over 12.5%. As an application example, we applied our model to the Stanford Question Answering Dataset to show ontology generation in a real field. The results showed that our model can generate a good ontology, with some exceptions in the real field, indicating future research directions to improve the quality.
引用
收藏
页数:16
相关论文
共 50 条
  • [41] Syntax-Enhanced Pretrained Language Models for Aspect-Level Sentiment Classification
    Yuan, Li
    Wang, Jin
    Lee, Lung-Hao
    Zhang, Xuejie
    IEEE TRANSACTIONS ON COMPUTATIONAL SOCIAL SYSTEMS, 2024,
  • [42] Injecting User Identity Into Pretrained Language Models for Document-Level Sentiment Classification
    Cao, Xinlei
    Yu, Jinyang
    Zhuang, Yan
    IEEE ACCESS, 2022, 10 : 30157 - 30167
  • [43] Layerwised multimodal knowledge distillation for vision-language pretrained model
    Wang, Jin
    Liao, Dawei
    Zhang, You
    Xu, Dan
    Zhang, Xuejie
    NEURAL NETWORKS, 2024, 175
  • [44] KBioXLM: A Knowledge-anchored Biomedical Multilingual Pretrained Language Model
    Geng, Lei
    Yan, Xu
    Cao, Ziqiang
    Li, Juntao
    Li, Wenjie
    Li, Sujian
    Zhou, Xinjie
    Yang, Yang
    Zhang, Jun
    FINDINGS OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS (EMNLP 2023), 2023, : 11239 - 11250
  • [45] Comparative Study of Multiclass Text Classification in Research Proposals Using Pretrained Language Models
    Lee, Eunchan
    Lee, Changhyeon
    Ahn, Sangtae
    APPLIED SCIENCES-BASEL, 2022, 12 (09):
  • [46] Unsupervised Domain Adaptation of a Pretrained Cross-Lingual Language Model
    Li, Juntao
    He, Ruidan
    Ye, Hai
    Ng, Hwee Tou
    Bing, Lidong
    Yan, Rui
    PROCEEDINGS OF THE TWENTY-NINTH INTERNATIONAL JOINT CONFERENCE ON ARTIFICIAL INTELLIGENCE, 2020, : 3672 - 3678
  • [47] Pretrained domain-specific language model for natural language processing tasks in the AEC domain
    Zheng, Zhe
    Lu, Xin-Zheng
    Chen, Ke-Yin
    Zhou, Yu-Cheng
    Lin, Jia-Rui
    COMPUTERS IN INDUSTRY, 2022, 142
  • [48] Exploring Cross-Domain Pretrained Model for Hyperspectral Image Classification
    Lee, Hyungtae
    Eum, Sungmin
    Kwon, Heesung
    IEEE TRANSACTIONS ON GEOSCIENCE AND REMOTE SENSING, 2022, 60
  • [49] Automatic Component Prediction for Issue Reports Using Fine-Tuned Pretrained Language Models
    Wang, Dae-Sung
    Lee, Chan-Gun
    IEEE ACCESS, 2022, 10 : 131456 - 131468
  • [50] Chinese Prosodic Structure Prediction Based on a Pretrained Language Representation Model
    Zhang P.
    Lu C.
    Wang R.
    Zhang, Pengyuan (zhangpengyuan@hccl.ioa.ac.cn), 1600, Tianjin University (53): : 265 - 271