ALSI-Transformer: Transformer-Based Code Comment Generation With Aligned Lexical and Syntactic Information

被引:1
|
作者
Park, Youngmi [1 ]
Park, Ahjeong [1 ]
Kim, Chulyun [1 ]
机构
[1] Sookmyung Womens Univ, Dept Informat Technol Engn, Seoul 04310, South Korea
关键词
Codes; Source coding; Syntactics; Data mining; Transformers; Machine translation; Logic gates; Program comprehension; comment generation; natural language processing; deep learning;
D O I
10.1109/ACCESS.2023.3268638
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
Code comments explain the operational process of a computer program and increase the long-term productivity of programming tasks such as debugging and maintenance. Therefore, developing methods that automatically generate natural language comments from programming code is required. With the development of deep learning, various excellent models in the natural language processing domain have been applied for comment generation tasks, and recent studies have improved performance by simultaneously using the lexical information of the code token and the syntactical information obtained from the syntax tree. In this paper, to improve the accuracy of automatic comment generation, we introduce a novel syntactic sequence, Code-Aligned Type sequence (CAT), to align the order and length of lexical and syntactic information, and we propose a new neural network model, Aligned Lexical and Syntactic information-Transformer (ALSI-Transformer), based on a transformer that encodes the aligned multi-modal information with convolution and embedding aggregation layers. Through in-depth experiments, we compared ALSI-Transformer with current baseline methods using standard machine translation metrics and demonstrate that the proposed method achieves state-of-the-art performance in code comment generation.
引用
收藏
页码:39037 / 39047
页数:11
相关论文
共 50 条
  • [1] SeTransformer: A Transformer-Based Code Semantic Parser for Code Comment Generation
    Li, Zheng
    Wu, Yonghao
    Peng, Bin
    Chen, Xiang
    Sun, Zeyu
    Liu, Yong
    Paul, Doyle
    IEEE TRANSACTIONS ON RELIABILITY, 2023, 72 (01) : 258 - 273
  • [2] An Empirical Study of Code Smells in Transformer-based Code Generation Techniques
    Siddiq, Mohammed Latif
    Majumder, Shafayat H.
    Mim, Maisha R.
    Jajodia, Sourov
    Santos, Joanna C. S.
    2022 IEEE 22ND INTERNATIONAL WORKING CONFERENCE ON SOURCE CODE ANALYSIS AND MANIPULATION (SCAM 2022), 2022, : 71 - 82
  • [3] Multilingual Controllable Transformer-Based Lexical Simplification
    Sheang, Kim Cheng
    Saggion, Horacio
    PROCESAMIENTO DEL LENGUAJE NATURAL, 2023, (71): : 109 - 123
  • [4] Assessing the Syntactic Capabilities of Transformer-based Multilingual Language Models
    Perez-Mayos, Laura
    Taboas Garcia, Alba
    Mille, Simon
    Wanner, Leo
    FINDINGS OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS, ACL-IJCNLP 2021, 2021, : 3799 - 3812
  • [5] A Transformer-Based Variational Autoencoder for Sentence Generation
    Liu, Danyang
    Liu, Gongshen
    2019 INTERNATIONAL JOINT CONFERENCE ON NEURAL NETWORKS (IJCNN), 2019,
  • [6] Transformer-based Point Cloud Generation Network
    Xu, Rui
    Hui, Le
    Han, Yuehui
    Qian, Jianjun
    Xie, Jin
    PROCEEDINGS OF THE 31ST ACM INTERNATIONAL CONFERENCE ON MULTIMEDIA, MM 2023, 2023, : 4169 - 4177
  • [7] Transformer-based partner dance motion generation
    Wu, Ying
    Wu, Zizhao
    Ji, Chengtao
    Engineering Applications of Artificial Intelligence, 2025, 139
  • [8] TBCUP: A Transformer-based Code Comments Updating Approach
    Liu, Shifan
    Cui, Zhanqi
    Chen, Xiang
    Yang, Jun
    Li, Li
    Zheng, Liwei
    2023 IEEE 47TH ANNUAL COMPUTERS, SOFTWARE, AND APPLICATIONS CONFERENCE, COMPSAC, 2023, : 892 - 897
  • [9] Transformer-based code model with compressed hierarchy representation
    Kechi Zhang
    Jia Li
    Zhuo Li
    Zhi Jin
    Ge Li
    Empirical Software Engineering, 2025, 30 (2)
  • [10] Progressive Transformer-Based Generation of Radiology Reports
    Nooralahzadeh, Farhad
    Gonzalez, Nicolas Perez
    Frauenfelder, Thomas
    Fujimoto, Koji
    Krauthammer, Michael
    FINDINGS OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS, EMNLP 2021, 2021, : 2824 - 2832