A transformer-based neural network framework for full names prediction with abbreviations and contexts

被引:0
|
作者
Ye, Ziming [1 ,2 ]
Li, Shuangyin [1 ]
机构
[1] South China Normal Univ, Sch Comp Sci, Guangzhou, Guangdong, Peoples R China
[2] Shandong Univ, Sch Comp Sci & Technol, Shandong, Peoples R China
基金
中国国家自然科学基金;
关键词
Full name inference; Multi-attention mask; Various grained contexts; Abbreviation; MODEL;
D O I
10.1016/j.datak.2023.102275
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
With the rapid spread of information, abbreviations are used more and more common because they are convenient. However, the duplication of abbreviations can lead to confusion in many cases, such as information management and information retrieval. The resultant confusion annoys users. Thus, inferring a full name from an abbreviation has practical and significant advantages. The bulk of studies in the literature mainly inferred full names based on rule-based methods, statistical models, the similarity of representation, etc. However, these methods are unable to use various grained contexts properly. In this paper, we propose a flexible framework of Multi-attention mask Abbreviation Context and Full name language model, named MACF to address the problem. With the abbreviation and contexts as the inputs, the MACF can automatically predict a full name by generation, where the contexts can be variously grained. That is, different grained contexts ranging from coarse to fine can be selected to perform such complicated tasks in which contexts include paragraphs, several sentences, or even just a few keywords. A novel multi-attention mask mechanism is also proposed, which allows the model to learn the relationships among abbreviations, contexts, and full names, a process that makes the most of various grained contexts. The three corpora of different languages and fields were analyzed and measured with seven metrics in various aspects to evaluate the proposed framework. According to the experimental results, the MACF yielded more significant and consistent outputs than other baseline methods. Moreover, we discuss the significance and findings, and give the case studies to show the performance in real applications.
引用
下载
收藏
页数:20
相关论文
共 50 条
  • [41] Generalizability of Convolutional Neural Network and Vision Transformer-Based OCT Segmentation Models
    Pely, Adam
    Wu, Zhichao
    Leng, Theodore
    Gao, Simon S.
    Chen, Hao
    Hejrati, Mohsen
    Zhang, Miao
    INVESTIGATIVE OPHTHALMOLOGY & VISUAL SCIENCE, 2023, 64 (08)
  • [42] Progressive Guidance Categorization Using Transformer-Based Deep Neural Network Architecture
    Aurpa, Tanjim Taharat
    Ahmed, Md Shoaib
    Sadik, Rifat
    Anwar, Sabbir
    Adnan, Md Abdul Mazid
    Anwar, Md Musfique
    HYBRID INTELLIGENT SYSTEMS, HIS 2021, 2022, 420 : 344 - 353
  • [43] TR-Net: A Transformer-Based Neural Network for Point Cloud Processing
    Liu, Luyao
    Chen, Enqing
    Ding, Yingqiang
    MACHINES, 2022, 10 (07)
  • [44] A transformer-based framework for enterprise sales forecasting
    Sun, Yupeng
    Li, Tian
    PeerJ Computer Science, 2024, 10 : 1 - 14
  • [45] Korean Sign Language Recognition Using Transformer-Based Deep Neural Network
    Shin, Jungpil
    Musa Miah, Abu Saleh
    Hasan, Md. Al Mehedi
    Hirooka, Koki
    Suzuki, Kota
    Lee, Hyoun-Sup
    Jang, Si-Woong
    APPLIED SCIENCES-BASEL, 2023, 13 (05):
  • [46] A Transformer-Based Framework for Scene Text Recognition
    Selvam, Prabu
    Koilraj, Joseph Abraham Sundar
    Tavera Romero, Carlos Andres
    Alharbi, Meshal
    Mehbodniya, Abolfazl
    Webber, Julian L.
    Sengan, Sudhakar
    IEEE ACCESS, 2022, 10 : 100895 - 100910
  • [47] Fastformer: Transformer-Based Fast Reasoning Framework
    Zhu, Wenjuan
    Guo, Ling
    Zhang, Tianxiang
    Han, Feng
    Wei, Yi
    Gong, Xiaoqing
    Xu, Pengfei
    Guo, Jing
    FOURTEENTH INTERNATIONAL CONFERENCE ON GRAPHICS AND IMAGE PROCESSING, ICGIP 2022, 2022, 12705
  • [48] Time-Aware Transformer-based Network for Clinical Notes Series Prediction
    Zhang, Dongyu
    Thadajarassiri, Jidapa
    Sen, Cansu
    Rundensteiner, Elke
    MACHINE LEARNING FOR HEALTHCARE CONFERENCE, VOL 126, 2020, 126 : 566 - 587
  • [49] Transformer-based structural seismic response prediction
    Zhang, Qingyu
    Guo, Maozi
    Zhao, Lingling
    Li, Yang
    Zhang, Xinxin
    Han, Miao
    STRUCTURES, 2024, 61
  • [50] A Transformer-based Neural Architecture Search Method
    Wang, Shang
    Tang, Huanrong
    Ouyang, Jianquan
    PROCEEDINGS OF THE 2023 GENETIC AND EVOLUTIONARY COMPUTATION CONFERENCE COMPANION, GECCO 2023 COMPANION, 2023, : 691 - 694