On the Uses of Large Language Models to Design End-to-end Learning Semantic Communication

被引:0
|
作者
Wang, Ying [1 ]
Sun, Zhuo [1 ]
Fan, Jinpo [1 ]
Ma, Hao [1 ]
机构
[1] Beijing Univ Posts & Telecommun, Key Lab Univ Wireless Commun, Minist Educ, Beijing 100876, Peoples R China
基金
中国国家自然科学基金;
关键词
end-to-end learning; large language models; semantic communication;
D O I
10.1109/WCNC57260.2024.10570717
中图分类号
TP3 [计算技术、计算机技术];
学科分类号
0812 ;
摘要
Deep learning-based semantic communication is a promising research direction for next-generation communication systems. The emergence of large language models(LLMs) with remarkable semantic comprehension abilities leads us to consider whether LLMs can be used in semantic communication to enhance model's performance. In this paper, we discuss the main implementing details of the idea by proposing a general end-to-end learning semantic communication model with LLM, including subword-level tokenization, a rate adapter based on gradients for matching the rate requirements of any channel codec and fine-tuning for possessing private background knowledge. By taking Bidirectional and Auto-Regressive Transformers (BART) and Generative Pre-trained Transformer 2 (GPT2) as examples, we demonstrate how we can utilize various structures of LLMs to design semantic codecs. In terms of semantic fidelity, generalizability to cross-scenario, and complexity, results reveal that the LLM-based semantic communication system achieves exciting performance. We hope this initial work can inspire more research devoted to this field.
引用
收藏
页数:6
相关论文
共 50 条
  • [21] Building Semantic Communication System via Molecules: An End-to-End Training Approach
    Cheng Yukun
    Chen Wei
    Ai Bo
    ChinaCommunications, 2024, 21 (07) : 113 - 124
  • [22] Deep Learning in Physical Layer: Review on Data Driven End-to-End Communication Systems and Their Enabling Semantic Applications
    Islam, Nazmul
    Shin, Seokjoo
    IEEE OPEN JOURNAL OF THE COMMUNICATIONS SOCIETY, 2024, 5 : 4207 - 4240
  • [23] KNOWLEDGE TRANSFER FROM LARGE-SCALE PRETRAINED LANGUAGE MODELS TO END-TO-END SPEECH RECOGNIZERS
    Kubo, Yotaro
    Karita, Shigeki
    Bacchiani, Michiel
    2022 IEEE INTERNATIONAL CONFERENCE ON ACOUSTICS, SPEECH AND SIGNAL PROCESSING (ICASSP), 2022, : 8512 - 8516
  • [24] Online Continual Learning of End-to-End Speech Recognition Models
    Yang, Muqiao
    Lane, Ian
    Watanabe, Shinji
    INTERSPEECH 2022, 2022, : 2668 - 2672
  • [25] IMPROVING END-TO-END MODELS FOR SET PREDICTION IN SPOKEN LANGUAGE UNDERSTANDING
    Kuo, Hong-Kwang J.
    Tuske, Zoltan
    Thomas, Samuel
    Kingsbury, Brian
    Saon, George
    2022 IEEE INTERNATIONAL CONFERENCE ON ACOUSTICS, SPEECH AND SIGNAL PROCESSING (ICASSP), 2022, : 7162 - 7166
  • [26] Recent progress in deep end-to-end models for spoken language processing
    Audhkhasi, K.
    Rosenberg, A.
    Saon, G.
    Sethy, A.
    Ramabhadran, B.
    Chen, S.
    Picheny, M.
    IBM JOURNAL OF RESEARCH AND DEVELOPMENT, 2017, 61 (4-5)
  • [27] End-to-end latency characterization of task communication models for automotive systems
    Martinez, Jorge
    Sanudo, Ignacio
    Bertogna, Marko
    REAL-TIME SYSTEMS, 2020, 56 (03) : 315 - 347
  • [28] End-to-end latency characterization of task communication models for automotive systems
    Jorge Martinez
    Ignacio Sañudo
    Marko Bertogna
    Real-Time Systems, 2020, 56 : 315 - 347
  • [29] End-to-End Phoneme Recognition using Models from Semantic Image Segmentation
    Gao, Wei
    Hashemi-Sakhtsari, Ahmad
    McDonnell, Mark D.
    2020 INTERNATIONAL JOINT CONFERENCE ON NEURAL NETWORKS (IJCNN), 2020,
  • [30] Defining the "to" in end-to-end models
    Mitra, Aditee
    Davis, Cabell
    PROGRESS IN OCEANOGRAPHY, 2010, 84 (1-2) : 39 - 42