API + Code = Better Code Summary? - Insights from an Exploratory Study
Automatic code summarization techniques aid in program comprehension as they try to generate a human-level summary in natural language from a programming language. Recent research in this area has seen significant developments from basic Seq2Seq models to different flavors of Transformer models, which try to encode the structural components of the source code using some input representation. Apart from the source code itself, other components, such as API knowledge, have previously been helpful in code summarization using recurrent neural networks (RNN) as it gives crucial information about the code’s functionality. So, in this article, along with the source code structure, we explore the importance of API knowledge in code summarization and try to understand whether it helps in improving the summaries. Our model uses a Transformer-based architecture containing two encoders for two input modules, source code and API sequences, and a joint decoder to generate summaries combining the outputs of two encoders. We experimented with our proposed model on a dataset of java projects collected from GitHub containing around 87K <Java Method, API Sequence, Comment> triplets. The experiments show our model outperforms most of the existing RNN-based approaches, but the overall performance does not improve compared with the state-of-the-art approach using Transformers. Thus, the results show that although API information is helpful for code summarization, we still need better methods to extract the valuable information from the API sequences.
Fri 18 NovDisplayed time zone: Beijing, Chongqing, Hong Kong, Urumqi change
16:00 - 17:30 | |||
16:00 20mResearch paper | Measuring design compliance using neural language models – an automotive case study PROMISE Dhasarathy Parthasarathy Volvo AB, Cecilia Ekelin Volvo AB, Anjali Karri Volvo AB, Jiapeng Sun Volvo AB, Panagiotis Moraitis Volvo AB | ||
16:20 20mResearch paper | API + Code = Better Code Summary? - Insights from an Exploratory Study PROMISE | ||
16:40 20mResearch paper | LOGI: An Empirical Model of Heat-Induced Disk Drive Data Loss and its Implications for Data Recovery PROMISE Hammad Ahmad University of Michigan, Colton Holoday , Ian Bertram University of Michigan, Kevin Angstadt , Zohreh Sharafi Polytechnique Montréal, Westley Weimer University of Michigan | ||
17:00 10mDay closing | Farewell PROMISE 2022 PROMISE Shane McIntosh University of Waterloo, Weiyi Shang Concordia University, Gema Rodríguez-Pérez University of British Columbia (UBC) |