ALSI-Transformer: Transformer-Based Code Comment Generation with Aligned Lexical and Syntactic Information
Citations

WEB OF SCIENCE

5
Citations

SCOPUS

8

초록

Code comments explain the operational process of a computer program and increase the long-term productivity of programming tasks such as debugging and maintenance. Therefore, developing methods that automatically generate natural language comments from programming code is required. With the development of deep learning, various excellent models in the natural language processing domain have been applied for comment generation tasks, and recent studies have improved performance by simultaneously using the lexical information of the code token and the syntactical information obtained from the syntax tree. In this paper, to improve the accuracy of automatic comment generation, we introduce a novel syntactic sequence, Code-Aligned Type sequence (CAT), to align the order and length of lexical and syntactic information, and we propose a new neural network model, Aligned Lexical and Syntactic information-Transformer (ALSI-Transformer), based on a transformer that encodes the aligned multi-modal information with convolution and embedding aggregation layers. Through in-depth experiments, we compared ALSI-Transformer with current baseline methods using standard machine translation metrics and demonstrate that the proposed method achieves state-of-the-art performance in code comment generation. Author

키워드

Codescomment generationData miningdeep learningLogic gatesMachine translationnatural language processingprogram comprehensionSource codingSyntacticsTransformers
제목
ALSI-Transformer: Transformer-Based Code Comment Generation with Aligned Lexical and Syntactic Information
저자
Park, YoungmiPark, AhjeongKim, Chulyun
DOI
10.1109/ACCESS.2023.3268638
발행일
2023-04
유형
Article in Press
저널명
IEEE Access
11
페이지
39037 ~ 39047