An Eclectic Approach for Enhancing Language Models Through Rich Embedding Features

被引:0
|
作者
Aldana-Bobadilla, Edwin [1 ,2 ]
Sosa-Sosa, Victor Jesus [2 ]
Molina-Villegas, Alejandro [1 ,3 ]
Gazca-Hernandez, Karina [2 ]
Olivas, Jose Angel [4 ]
机构
[1] CONAHCYT, Mexico City 03940, Mexico
[2] Cinvestav, Unidad Tamaulipas, Ciudad Victoria 87130, Tamaulipas, Mexico
[3] Ctr Invest Ciencias Invest Geoespacial, Mexico City 14240, Mexico
[4] Univ Castilla La Mancha, Grp SMILe, Ciudad Real 13071, Spain
来源
IEEE ACCESS | 2024年 / 12卷
关键词
Task analysis; Semantics; Transformers; Neurons; Linguistics; Self-organizing feature maps; Text analysis; Self-organizing map; word embeddings; feature extraction; natural language processing;
D O I
10.1109/ACCESS.2024.3422971
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
Text processing is a fundamental aspect of Natural Language Processing (NLP) and is crucial for various applications in fields such as artificial intelligence, data science, and information retrieval. It plays a core role in language models. Most text-processing approaches focus on describing and synthesizing, to a greater or lesser degree, lexical, syntactic, and semantic properties of text in the form of numerical vectors that induce a metric space, in which, it is possible to find underlying patterns and structures related to the original text. Since each approach has strengths and weaknesses, finding a single approach that perfectly extracts representative text properties for every task and application domain is hard. This paper proposes a novel approach capable of synthesizing information from heterogeneous state-of-the-art text processing approaches into a unified representation. Encouraging results demonstrate that using this representation in popular machine-learning tasks not only leads to superior performance but also offers notable advantages in memory efficiency and preservation of underlying information of the distinct sources involved in such a representation.
引用
收藏
页码:100921 / 100938
页数:18
相关论文
共 50 条
  • [1] Enhancing performance of transformer-based models in natural language understanding through word importance embedding
    Hong, Seung-Kyu
    Jang, Jae-Seok
    Kwon, Hyuk-Yoon
    KNOWLEDGE-BASED SYSTEMS, 2024, 304
  • [2] Floating-Point Embedding: Enhancing the Mathematical Comprehension of Large Language Models
    Jin, Xiaoxiao
    Mao, Chenyang
    Yue, Dengfeng
    Leng, Tuo
    SYMMETRY-BASEL, 2024, 16 (04):
  • [3] Embedding Recycling for Language Models
    Saad-Falcon, Jon
    Singh, Amanpreet
    Soldaini, Luca
    D'Arcy, Mike
    Cohan, Arman
    Downey, Doug
    17TH CONFERENCE OF THE EUROPEAN CHAPTER OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS, EACL 2023, 2023, : 1933 - 1953
  • [4] Enhancing Multilingual Capabilities of Large Language Models through Self-Distillation from Resource-Rich Languages
    Zhang, Yuanchi
    Wang, Yile
    Liu, Zijun
    Wang, Shuo
    Wang, Xiaolong
    Li, Peng
    Sun, Maosong
    Liu, Yang
    PROCEEDINGS OF THE 62ND ANNUAL MEETING OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS, VOL 1: LONG PAPERS, 2024, : 11189 - 11204
  • [5] Enhancing healthcare resource allocation through large language models
    Wan, Fang
    Wang, Kezhi
    Wang, Tao
    Qin, Hu
    Fondrevelle, Julien
    Duclos, Antoine
    SWARM AND EVOLUTIONARY COMPUTATION, 2025, 94
  • [6] Enhancing Large Language Models Through External Domain Knowledge
    Welz, Laslo
    Lanquillon, Carsten
    ARTIFICIAL INTELLIGENCE IN HCI, PT III, AI-HCI 2024, 2024, 14736 : 135 - 146
  • [7] Enhancing Supply Chain Efficiency through Retrieve-Augmented Generation Approach in Large Language Models
    Zhu, Beilei
    Vuppalapati, Chandrasekar
    2024 IEEE 10TH INTERNATIONAL CONFERENCE ON BIG DATA COMPUTING SERVICE AND MACHINE LEARNING APPLICATIONS, BIGDATASERVICE 2024, 2024, : 117 - 121
  • [8] Enhancing health assessments with large language models: A methodological approach
    Wang, Xi
    Zhou, Yujia
    Zhou, Guangyu
    APPLIED PSYCHOLOGY-HEALTH AND WELL BEING, 2025, 17 (01)
  • [9] Enhancing Interaction Flow Modeling Language Metamodels for Designing Features of Rich Internet Applications
    Wakil, Karzan
    Jawawi, Dayang N. A.
    Rachmat, Haris
    INTERNATIONAL JOURNAL OF INTEGRATED ENGINEERING, 2018, 10 (06): : 97 - 105
  • [10] Enhancing Bias Assessment for Complex Term Groups in Language Embedding Models: Quantitative Comparison of Methods
    Gray, Magnus
    Milanova, Mariofanna
    Wu, Leihong
    JMIR MEDICAL INFORMATICS, 2024, 12