A transformer-based neural network framework for full names prediction with abbreviations and contexts

被引:0
|
作者
Ye, Ziming [1 ,2 ]
Li, Shuangyin [1 ]
机构
[1] South China Normal Univ, Sch Comp Sci, Guangzhou, Guangdong, Peoples R China
[2] Shandong Univ, Sch Comp Sci & Technol, Shandong, Peoples R China
基金
中国国家自然科学基金;
关键词
Full name inference; Multi-attention mask; Various grained contexts; Abbreviation; MODEL;
D O I
10.1016/j.datak.2023.102275
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
With the rapid spread of information, abbreviations are used more and more common because they are convenient. However, the duplication of abbreviations can lead to confusion in many cases, such as information management and information retrieval. The resultant confusion annoys users. Thus, inferring a full name from an abbreviation has practical and significant advantages. The bulk of studies in the literature mainly inferred full names based on rule-based methods, statistical models, the similarity of representation, etc. However, these methods are unable to use various grained contexts properly. In this paper, we propose a flexible framework of Multi-attention mask Abbreviation Context and Full name language model, named MACF to address the problem. With the abbreviation and contexts as the inputs, the MACF can automatically predict a full name by generation, where the contexts can be variously grained. That is, different grained contexts ranging from coarse to fine can be selected to perform such complicated tasks in which contexts include paragraphs, several sentences, or even just a few keywords. A novel multi-attention mask mechanism is also proposed, which allows the model to learn the relationships among abbreviations, contexts, and full names, a process that makes the most of various grained contexts. The three corpora of different languages and fields were analyzed and measured with seven metrics in various aspects to evaluate the proposed framework. According to the experimental results, the MACF yielded more significant and consistent outputs than other baseline methods. Moreover, we discuss the significance and findings, and give the case studies to show the performance in real applications.
引用
下载
收藏
页数:20
相关论文
共 50 条
  • [21] A Transformer-Based Network for Full Object Pose Estimation with Depth Refinement
    Abdulsalam, Mahmoud
    Ahiska, Kenan
    Aouf, Nabil
    ADVANCED INTELLIGENT SYSTEMS, 2024,
  • [22] Transformer-Based Unified Neural Network for Quality Estimation and Transformer-Based Re-decoding Model for Machine Translation
    Chen, Cong
    Zong, Qinqin
    Luo, Qi
    Qiu, Bailian
    Li, Maoxi
    MACHINE TRANSLATION, CCMT 2020, 2020, 1328 : 66 - 75
  • [23] Automated redaction of names in adverse event reports using transformer-based neural networks
    Eva-Lisa Meldau
    Shachi Bista
    Carlos Melgarejo-González
    G. Niklas Norén
    BMC Medical Informatics and Decision Making, 24 (1)
  • [24] A Transformer-Based Neural Network for Gait Prediction in Lower Limb Exoskeleton Robots Using Plantar Force
    Ren, Jiale
    Wang, Aihui
    Li, Hengyi
    Yue, Xuebin
    Meng, Lin
    SENSORS, 2023, 23 (14)
  • [25] HyperTeNet: Hypergraph and Transformer-based Neural Network for Personalized List Continuation
    Vijaikumar, M.
    Hada, Deepesh
    Shevade, Shirish
    2021 21ST IEEE INTERNATIONAL CONFERENCE ON DATA MINING (ICDM 2021), 2021, : 1210 - 1215
  • [26] SolarDetector: A Transformer-based Neural Network for the Detection and Masking of Solar Panels
    Salama, Abdulrahman
    Hendawi, Abdeltawab
    Franklin, Richard
    Al-Masri, Eyhab
    Deshpande, Anish
    Ali, Mohamed
    31ST ACM SIGSPATIAL INTERNATIONAL CONFERENCE ON ADVANCES IN GEOGRAPHIC INFORMATION SYSTEMS, ACM SIGSPATIAL GIS 2023, 2023, : 610 - 613
  • [27] A Context-Integrated Transformer-Based Neural Network for Auction Design
    Duan, Zhijian
    Tang, Jingwu
    Yin, Yutong
    Feng, Zhe
    Yan, Xiang
    Zaheer, Manzil
    Deng, Xiaotie
    INTERNATIONAL CONFERENCE ON MACHINE LEARNING, VOL 162, 2022,
  • [28] A novel transformer-based neural network model for tool wear estimation
    Liu, Hui
    Liu, Zhenyu
    Jia, Weiqiang
    Lin, Xianke
    Zhang, Shuo
    MEASUREMENT SCIENCE AND TECHNOLOGY, 2020, 31 (06)
  • [29] Learning continuous dynamic network representation with transformer-based temporal graph neural network
    Li, Yingji
    Wu, Yue
    Sun, Mingchen
    Yang, Bo
    Wang, Ying
    INFORMATION SCIENCES, 2023, 649
  • [30] TF-Predictor: Transformer-Based Prerouting Path Delay Prediction Framework
    Cao, Peng
    He, Guoqing
    Yang, Tai
    IEEE TRANSACTIONS ON COMPUTER-AIDED DESIGN OF INTEGRATED CIRCUITS AND SYSTEMS, 2023, 42 (07) : 2227 - 2237