A self-attention hybrid emoji prediction model for code-mixed language: (Hinglish)

被引:0
|
作者
Gadde Satya Sai Naga Himabindu
Rajat Rao
Divyashikha Sethia
机构
[1] Delhi Technological University,Department of Computer Engineering
来源
关键词
Emoji prediction; Hinglish; Code mixed; Deep learning; Hybrid model;
D O I
暂无
中图分类号
学科分类号
摘要
Emojis are an essential tool for communication, and various resource-rich languages such as English use emoji prediction systems. However, there is limited research on emoji prediction for resource-poor and code-mixed languages such as Hinglish (Hindi + English), the fourth most used code-mixed language globally. This paper proposes a novel Hinglish Emoji Prediction (HEP) dataset created using Twitter as a corpus and a hybrid emoji prediction model BiLSTM attention random forest (BARF) for code-mixed Hinglish language. The proposed BARF model combines deep learning features with machine learning classification. It begins with BiLSTM to capture the context and then proceeds to self-attention to extract significant texts. Finally, it uses random forest to categorize the features to predict an emoji. The self-attention mechanism aids learning since Hinglish, a code-mixed language, lacks proper grammatical rules. The combination of deep learning and machine learning algorithms and attention is novel to emoji prediction in the code-mixed language(Hinglish). Results on the HEP dataset indicate that the BARF model outperformed previous multilingual and baseline emoji prediction models. It achieved an accuracy of 61.14%, precision of 0.66, recall of 0.59, and F1 score of 0.59.
引用
收藏
相关论文
共 50 条
  • [31] HTNet: A Hybrid Model Boosted by Triple Self-attention for Crowd Counting
    Li, Yang
    Yin, Baoqun
    PATTERN RECOGNITION AND COMPUTER VISION, PRCV 2023, PT XII, 2024, 14436 : 290 - 301
  • [32] A Comprehensive Understanding of Code-Mixed Language Semantics Using Hierarchical Transformer
    Suresh, Tharun
    Sengupta, Ayan
    Akhtar, Md Shad
    Chakraborty, Tanmoy
    IEEE TRANSACTIONS ON COMPUTATIONAL SOCIAL SYSTEMS, 2024, 11 (03) : 4139 - 4148
  • [33] Automatic Text Summarization for Code-Mixed Language using Fuzzy Logic
    Tayal, Madhuri A.
    Tayal, Animesh
    Kokardekar, Pratibha
    BIOSCIENCE BIOTECHNOLOGY RESEARCH COMMUNICATIONS, 2020, 13 (14): : 150 - 155
  • [34] Self-Attention and Dynamic Convolution Hybrid Model for Neural Machine Translation
    Zhang, Zhebin
    Wu, Sai
    Chen, Gang
    Jiang, Dawei
    11TH IEEE INTERNATIONAL CONFERENCE ON KNOWLEDGE GRAPH (ICKG 2020), 2020, : 352 - 359
  • [35] Multilevel Self-Attention Model and its Use on Medical Risk Prediction
    Zeng, Xianlong
    Feng, Yunyi
    Moosavinasab, Soheil
    Lin, Deborah
    Lin, Simon
    Liu, Chang
    PACIFIC SYMPOSIUM ON BIOCOMPUTING 2020, 2020, : 115 - 126
  • [36] Session interest model for CTR prediction based on self-attention mechanism
    Qianqian Wang
    Fang’ai Liu
    Xiaohui Zhao
    Qiaoqiao Tan
    Scientific Reports, 12
  • [37] A Self-Attention Integrated Learning Model for Landing Gear Performance Prediction
    Lin, Lin
    Tong, Changsheng
    Guo, Feng
    Fu, Song
    Lv, Yancheng
    He, Wenhui
    SENSORS, 2023, 23 (13)
  • [38] Cascade Prediction model based on Dynamic Graph Representation and Self-Attention
    Zhang F.
    Wang X.
    Wang R.
    Tang Q.
    Han Y.
    Dianzi Keji Daxue Xuebao/Journal of the University of Electronic Science and Technology of China, 2022, 51 (01): : 83 - 90
  • [39] A Self-Attention Model for Next Location Prediction Based on Semantic Mining
    Lu, Eric Hsueh-Chan
    Lin, You-Ru
    ISPRS INTERNATIONAL JOURNAL OF GEO-INFORMATION, 2023, 12 (10)
  • [40] Solar irradiance prediction based on self-attention recursive model network
    Kang, Ting
    Wang, Huaizhi
    Wu, Ting
    Peng, Jianchun
    Jiang, Hui
    FRONTIERS IN ENERGY RESEARCH, 2022, 10