On the Uses of Large Language Models to Design End-to-end Learning Semantic Communication

被引:0
|
作者
Wang, Ying [1 ]
Sun, Zhuo [1 ]
Fan, Jinpo [1 ]
Ma, Hao [1 ]
机构
[1] Beijing Univ Posts & Telecommun, Key Lab Univ Wireless Commun, Minist Educ, Beijing 100876, Peoples R China
基金
中国国家自然科学基金;
关键词
end-to-end learning; large language models; semantic communication;
D O I
10.1109/WCNC57260.2024.10570717
中图分类号
TP3 [计算技术、计算机技术];
学科分类号
0812 ;
摘要
Deep learning-based semantic communication is a promising research direction for next-generation communication systems. The emergence of large language models(LLMs) with remarkable semantic comprehension abilities leads us to consider whether LLMs can be used in semantic communication to enhance model's performance. In this paper, we discuss the main implementing details of the idea by proposing a general end-to-end learning semantic communication model with LLM, including subword-level tokenization, a rate adapter based on gradients for matching the rate requirements of any channel codec and fine-tuning for possessing private background knowledge. By taking Bidirectional and Auto-Regressive Transformers (BART) and Generative Pre-trained Transformer 2 (GPT2) as examples, we demonstrate how we can utilize various structures of LLMs to design semantic codecs. In terms of semantic fidelity, generalizability to cross-scenario, and complexity, results reveal that the LLM-based semantic communication system achieves exciting performance. We hope this initial work can inspire more research devoted to this field.
引用
收藏
页数:6
相关论文
共 50 条
  • [1] Semantic Complexity in End-to-End Spoken Language Understanding
    McKenna, Joseph P.
    Choudhary, Samridhi
    Saxon, Michael
    Strimel, Grant P.
    Mouchtaris, Athanasios
    INTERSPEECH 2020, 2020, : 4273 - 4277
  • [2] End-to-End Learning for Integrated Sensing and Communication
    Mateos-Ramos, Jose Miguel
    Song, Jinxiang
    Wu, Yibo
    Hager, Christian
    Keskin, Musa Furkan
    Yajnanarayana, Vijaya
    Wymeersch, Henk
    IEEE INTERNATIONAL CONFERENCE ON COMMUNICATIONS (ICC 2022), 2022, : 1942 - 1947
  • [3] Learning Neural Models for End-to-End Clustering
    Meier, Benjamin Bruno
    Elezi, Ismail
    Amirian, Mohammadreza
    Duerr, Oliver
    Stadelmann, Thilo
    ARTIFICIAL NEURAL NETWORKS IN PATTERN RECOGNITION, ANNPR 2018, 2018, 11081 : 126 - 138
  • [4] An automatic end-to-end chemical synthesis development platform powered by large language models
    Yixiang Ruan
    Chenyin Lu
    Ning Xu
    Yuchen He
    Yixin Chen
    Jian Zhang
    Jun Xuan
    Jianzhang Pan
    Qun Fang
    Hanyu Gao
    Xiaodong Shen
    Ning Ye
    Qiang Zhang
    Yiming Mo
    Nature Communications, 15 (1)
  • [5] INSIGHTS INTO END-TO-END LEARNING SCHEME FOR LANGUAGE IDENTIFICATION
    Cai, Weicheng
    Cai, Zexin
    Liu, Wenbo
    Wang, Xiaoqi
    Li, Ming
    2018 IEEE INTERNATIONAL CONFERENCE ON ACOUSTICS, SPEECH AND SIGNAL PROCESSING (ICASSP), 2018, : 5209 - 5213
  • [6] Design of network transmission model for end-to-end communication
    Chun, You
    PROCEEDINGS OF 2021 2ND INTERNATIONAL CONFERENCE ON ARTIFICIAL INTELLIGENCE AND INFORMATION SYSTEMS (ICAIIS '21), 2021,
  • [7] Effective End-to-End Vision Language Pretraining With Semantic Visual Loss
    Yang, Xiaofeng
    Liu, Fayao
    Lin, Guosheng
    IEEE TRANSACTIONS ON MULTIMEDIA, 2023, 25 : 8408 - 8417
  • [8] End-to-end Learning of Driving Models from Large-scale Video Datasets
    Xu, Huazhe
    Gao, Yang
    Yu, Fisher
    Darrell, Trevor
    30TH IEEE CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR 2017), 2017, : 3530 - 3538
  • [9] END-TO-END LEARNING OF PARSING MODELS FOR INFORMATION RETRIEVAL
    Gillenwater, Jennifer
    He, Xiaodong
    Gao, Jianfeng
    Deng, Li
    2013 IEEE INTERNATIONAL CONFERENCE ON ACOUSTICS, SPEECH AND SIGNAL PROCESSING (ICASSP), 2013, : 3312 - 3316
  • [10] Learning Diverse Models for End-to-End Ensemble Tracking
    Wang, Ning
    Zhou, Wengang
    Li, Houqiang
    IEEE Transactions on Image Processing, 2021, 30 : 2220 - 2231