A transformer-based neural network framework for full names prediction with abbreviations and contexts

被引:0
|
作者
Ye, Ziming [1 ,2 ]
Li, Shuangyin [1 ]
机构
[1] South China Normal Univ, Sch Comp Sci, Guangzhou, Guangdong, Peoples R China
[2] Shandong Univ, Sch Comp Sci & Technol, Shandong, Peoples R China
基金
中国国家自然科学基金;
关键词
Full name inference; Multi-attention mask; Various grained contexts; Abbreviation; MODEL;
D O I
10.1016/j.datak.2023.102275
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
With the rapid spread of information, abbreviations are used more and more common because they are convenient. However, the duplication of abbreviations can lead to confusion in many cases, such as information management and information retrieval. The resultant confusion annoys users. Thus, inferring a full name from an abbreviation has practical and significant advantages. The bulk of studies in the literature mainly inferred full names based on rule-based methods, statistical models, the similarity of representation, etc. However, these methods are unable to use various grained contexts properly. In this paper, we propose a flexible framework of Multi-attention mask Abbreviation Context and Full name language model, named MACF to address the problem. With the abbreviation and contexts as the inputs, the MACF can automatically predict a full name by generation, where the contexts can be variously grained. That is, different grained contexts ranging from coarse to fine can be selected to perform such complicated tasks in which contexts include paragraphs, several sentences, or even just a few keywords. A novel multi-attention mask mechanism is also proposed, which allows the model to learn the relationships among abbreviations, contexts, and full names, a process that makes the most of various grained contexts. The three corpora of different languages and fields were analyzed and measured with seven metrics in various aspects to evaluate the proposed framework. According to the experimental results, the MACF yielded more significant and consistent outputs than other baseline methods. Moreover, we discuss the significance and findings, and give the case studies to show the performance in real applications.
引用
下载
收藏
页数:20
相关论文
共 50 条
  • [31] A Transformer-Based Ensemble Framework for the Prediction of Protein-Protein Interaction Sites
    Mou, Minjie
    Pan, Ziqi
    Zhou, Zhimeng
    Zheng, Lingyan
    Zhang, Hanyu
    Shi, Shuiyang
    Li, Fengcheng
    Sun, Xiuna
    Zhu, Feng
    RESEARCH, 2023, 6
  • [32] TXtreme: transformer-based extreme value prediction framework for time series forecasting
    Hemant Yadav
    Amit Thakkar
    Discover Applied Sciences, 7 (2)
  • [33] Remaining Useful Life Prediction of Lithium-Ion Batteries by Using a Denoising Transformer-Based Neural Network
    Han, Yunlong
    Li, Conghui
    Zheng, Linfeng
    Lei, Gang
    Li, Li
    ENERGIES, 2023, 16 (17)
  • [34] Novel Transformer-based deep neural network for the prediction of post-refracturing production from oil wells
    Jia, Jing
    Li, Diquan
    Wang, Lichang
    Fan, Qinghu
    ADVANCES IN GEO-ENERGY RESEARCH, 2024, 13 (02): : 119 - 131
  • [35] VDGCNeT: A novel network-wide Virtual Dynamic Graph Convolution Neural network and Transformer-based traffic prediction model
    Zheng, Ge
    Chai, Wei Koong
    Zhang, Jiankang
    Katos, Vasilis
    KNOWLEDGE-BASED SYSTEMS, 2023, 275
  • [36] A transformer-based network for speech recognition
    Tang L.
    International Journal of Speech Technology, 2023, 26 (02) : 531 - 539
  • [37] RPConvformer: A novel Transformer-based deep neural networks for traffic flow prediction
    Wen, Yanjie
    Xu, Ping
    Li, Zhihong
    Xu, Wangtu
    Wang, Xiaoyu
    EXPERT SYSTEMS WITH APPLICATIONS, 2023, 218
  • [38] Carbon emissions forecasting based on temporal graph transformer-based attentional neural network
    Wu, Xingping
    Yuan, Qiheng
    Zhou, Chunlei
    Chen, Xiang
    Xuan, Donghai
    Song, Jinwei
    JOURNAL OF COMPUTATIONAL METHODS IN SCIENCES AND ENGINEERING, 2024, 24 (03) : 1405 - 1421
  • [39] A Transformer-Based Framework for Tiny Object Detection
    Liao, Yi-Kai
    Lin, Gong-Si
    Yeh, Mei-Chen
    2023 ASIA PACIFIC SIGNAL AND INFORMATION PROCESSING ASSOCIATION ANNUAL SUMMIT AND CONFERENCE, APSIPA ASC, 2023, : 373 - 377
  • [40] Empirical assessment of transformer-based neural network architecture in forecasting pollution trends
    Nath, Pritthijit
    Middya, Asif Iqbal
    Roy, Sarbani
    INTERNATIONAL JOURNAL OF DATA SCIENCE AND ANALYTICS, 2023,