CLG: Automated checklist generation for improved pull request quality

IF 7.5 1区 计算机科学 Q1 COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE
Expert Systems with Applications Pub Date : 2025-06-05 Epub Date: 2025-03-14 DOI:10.1016/j.eswa.2025.127178
Shuotong Bai, Chenkun Meng, Guodong Li, Huaxiao Liu, Lei Liu
{"title":"CLG: Automated checklist generation for improved pull request quality","authors":"Shuotong Bai,&nbsp;Chenkun Meng,&nbsp;Guodong Li,&nbsp;Huaxiao Liu,&nbsp;Lei Liu","doi":"10.1016/j.eswa.2025.127178","DOIUrl":null,"url":null,"abstract":"<div><div>The Pull-Based development model, widely embraced in open-source software (OSS), leverages closer global collaboration. However, the increasing number of contributors introduces challenges, particularly in PR quality. Mature repositories address this by using checklists in Pull Request Templates (PRTs). Despite their benefits, a mere 14.15% of popular GitHub repositories implement such checklists. To address this gap, we propose CLG, a Check-List Generation approach, utilizing techniques with a multi-label classifier and summary generation to automatically generate checklists from contributing guidelines. Evaluation results demonstrate CLG’s superiority in each sub-task. As for categorize paragraphs of contributing guidelines, CLG improves 9.81%, 8.44%, 11.84%, and 10.17% across metrics of Accuracy, Precision, Recall, and F1-score. In the task of note generation, CLG improves 16.33%, 15.80%, and 6.43% in terms of ROUGE metrics. To investigate our generated checklists from actual perspectives, we submit our results to the open-source community. The results show that our approach can help the majority of repository maintainers and the open-source community in managing and submitting their received PRs.</div></div>","PeriodicalId":50461,"journal":{"name":"Expert Systems with Applications","volume":"277 ","pages":"Article 127178"},"PeriodicalIF":7.5000,"publicationDate":"2025-06-05","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"Expert Systems with Applications","FirstCategoryId":"94","ListUrlMain":"https://www.sciencedirect.com/science/article/pii/S0957417425008000","RegionNum":1,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"2025/3/14 0:00:00","PubModel":"Epub","JCR":"Q1","JCRName":"COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE","Score":null,"Total":0}
引用次数: 0

Abstract

The Pull-Based development model, widely embraced in open-source software (OSS), leverages closer global collaboration. However, the increasing number of contributors introduces challenges, particularly in PR quality. Mature repositories address this by using checklists in Pull Request Templates (PRTs). Despite their benefits, a mere 14.15% of popular GitHub repositories implement such checklists. To address this gap, we propose CLG, a Check-List Generation approach, utilizing techniques with a multi-label classifier and summary generation to automatically generate checklists from contributing guidelines. Evaluation results demonstrate CLG’s superiority in each sub-task. As for categorize paragraphs of contributing guidelines, CLG improves 9.81%, 8.44%, 11.84%, and 10.17% across metrics of Accuracy, Precision, Recall, and F1-score. In the task of note generation, CLG improves 16.33%, 15.80%, and 6.43% in terms of ROUGE metrics. To investigate our generated checklists from actual perspectives, we submit our results to the open-source community. The results show that our approach can help the majority of repository maintainers and the open-source community in managing and submitting their received PRs.

Abstract Image

CLG:自动生成检查表以提高拉取请求的质量
在开源软件(OSS)中广泛采用的基于pull的开发模型利用了更紧密的全球协作。然而,越来越多的贡献者带来了挑战,特别是在PR质量方面。成熟的存储库通过使用拉请求模板(prt)中的检查表来解决这个问题。尽管有这些好处,但只有14.15%的流行GitHub存储库实现了这样的清单。为了解决这一差距,我们提出了CLG,一种检查表生成方法,利用多标签分类器和摘要生成技术从贡献指南自动生成检查表。评价结果显示了CLG在各子任务上的优势。对于贡献指南的分类段落,CLG在准确性、精密度、召回率和f1得分指标上分别提高了9.81%、8.44%、11.84%和10.17%。在生成注释的任务中,CLG在ROUGE指标方面分别提高了16.33%、15.80%和6.43%。为了从实际的角度调查我们生成的检查表,我们将结果提交给开源社区。结果表明,我们的方法可以帮助大多数存储库维护者和开源社区管理和提交他们收到的pr。
本文章由计算机程序翻译,如有差异,请以英文原文为准。
求助全文
约1分钟内获得全文 求助全文
来源期刊
Expert Systems with Applications
Expert Systems with Applications 工程技术-工程:电子与电气
CiteScore
13.80
自引率
10.60%
发文量
2045
审稿时长
8.7 months
期刊介绍: Expert Systems With Applications is an international journal dedicated to the exchange of information on expert and intelligent systems used globally in industry, government, and universities. The journal emphasizes original papers covering the design, development, testing, implementation, and management of these systems, offering practical guidelines. It spans various sectors such as finance, engineering, marketing, law, project management, information management, medicine, and more. The journal also welcomes papers on multi-agent systems, knowledge management, neural networks, knowledge discovery, data mining, and other related areas, excluding applications to military/defense systems.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
确定
请完成安全验证×
copy
已复制链接
快去分享给好友吧!
我知道了
右上角分享
点击右上角分享
0
联系我们:info@booksci.cn Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。 Copyright © 2023 布克学术 All rights reserved.
京ICP备2023020795号-1
ghs 京公网安备 11010802042870号
Book学术文献互助
Book学术文献互助群
群 号:604180095
Book学术官方微信
小红书