Vulnerability detection using BERT based LLM model with transparency obligation practice towards trustworthy AI

Jean Haurogné , Nihala Basheer , Shareeful Islam
{"title":"Vulnerability detection using BERT based LLM model with transparency obligation practice towards trustworthy AI","authors":"Jean Haurogné ,&nbsp;Nihala Basheer ,&nbsp;Shareeful Islam","doi":"10.1016/j.mlwa.2024.100598","DOIUrl":null,"url":null,"abstract":"<div><div>Vulnerabilities in the source code are one of the main causes of potential threats in software-intensive systems. There are a large number of vulnerabilities published each day, and effective vulnerability detection is critical to identifying and mitigating these vulnerabilities. AI has emerged as a promising solution to enhance vulnerability detection, offering the ability to analyse vast amounts of data and identify patterns indicative of potential threats. However, AI-based methods often face several challenges, specifically when dealing with large datasets and understanding the specific context of the problem. Large Language Model (LLM) is now widely considered to tackle more complex tasks and handle large datasets, which also exhibits limitations in terms of explaining the model outcome and existing works focus on providing overview of explainability and transparency. This research introduces a novel transparency obligation practice for vulnerability detection using BERT based LLMs. We address the black-box nature of LLMs by employing XAI techniques, unique combination of SHAP, LIME, heat map. We propose an architecture that combines the BERT model with transparency obligation practices, which ensures the assurance of transparency throughout the entire LLM life cycle. An experiment is performed with a large source code dataset to demonstrate the applicability of the proposed approach. The result shows higher accuracy of 91.8 % for the vulnerability detection and model explainability outcome is highly influenced by “vulnerable”, “function”, \"mysql_tmpdir_list\", “strmov” tokens using both SHAP and LIME framework. Heatmap of attention weights, highlights the local token interactions that aid in understanding the model's decision points.</div></div>","PeriodicalId":74093,"journal":{"name":"Machine learning with applications","volume":"18 ","pages":"Article 100598"},"PeriodicalIF":0.0000,"publicationDate":"2024-11-02","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"Machine learning with applications","FirstCategoryId":"1085","ListUrlMain":"https://www.sciencedirect.com/science/article/pii/S2666827024000744","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
引用次数: 0

Abstract

Vulnerabilities in the source code are one of the main causes of potential threats in software-intensive systems. There are a large number of vulnerabilities published each day, and effective vulnerability detection is critical to identifying and mitigating these vulnerabilities. AI has emerged as a promising solution to enhance vulnerability detection, offering the ability to analyse vast amounts of data and identify patterns indicative of potential threats. However, AI-based methods often face several challenges, specifically when dealing with large datasets and understanding the specific context of the problem. Large Language Model (LLM) is now widely considered to tackle more complex tasks and handle large datasets, which also exhibits limitations in terms of explaining the model outcome and existing works focus on providing overview of explainability and transparency. This research introduces a novel transparency obligation practice for vulnerability detection using BERT based LLMs. We address the black-box nature of LLMs by employing XAI techniques, unique combination of SHAP, LIME, heat map. We propose an architecture that combines the BERT model with transparency obligation practices, which ensures the assurance of transparency throughout the entire LLM life cycle. An experiment is performed with a large source code dataset to demonstrate the applicability of the proposed approach. The result shows higher accuracy of 91.8 % for the vulnerability detection and model explainability outcome is highly influenced by “vulnerable”, “function”, "mysql_tmpdir_list", “strmov” tokens using both SHAP and LIME framework. Heatmap of attention weights, highlights the local token interactions that aid in understanding the model's decision points.
使用基于 BERT 的 LLM 模型进行漏洞检测,履行透明义务,实现可信赖的人工智能
源代码中的漏洞是造成软件密集型系统潜在威胁的主要原因之一。每天都有大量漏洞发布,有效的漏洞检测对于识别和缓解这些漏洞至关重要。人工智能已成为加强漏洞检测的一种前景广阔的解决方案,它能够分析海量数据并识别表明潜在威胁的模式。然而,基于人工智能的方法往往面临一些挑战,特别是在处理大型数据集和了解问题的具体背景时。目前,人们普遍认为大型语言模型(LLM)可以解决更复杂的任务和处理大型数据集,但它在解释模型结果方面也表现出局限性,现有的工作主要集中在提供可解释性和透明度方面的概述。本研究介绍了使用基于 BERT 的 LLM 进行漏洞检测的新型透明度义务实践。我们采用 XAI 技术、SHAP、LIME 和热图的独特组合,解决了 LLM 的黑箱性质。我们提出了一种将 BERT 模型与透明度义务实践相结合的架构,可确保整个 LLM 生命周期的透明度。我们使用大型源代码数据集进行了实验,以证明所提方法的适用性。结果表明,使用 SHAP 和 LIME 框架,漏洞检测的准确率高达 91.8%,模型的可解释性结果受 "vulnerable"、"function"、"mysql_tmpdir_list "和 "strmov "令牌的影响很大。注意力权重热图,突出显示了有助于理解模型决策点的局部标记相互作用。
本文章由计算机程序翻译,如有差异,请以英文原文为准。
求助全文
约1分钟内获得全文 求助全文
来源期刊
Machine learning with applications
Machine learning with applications Management Science and Operations Research, Artificial Intelligence, Computer Science Applications
自引率
0.00%
发文量
0
审稿时长
98 days
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
确定
请完成安全验证×
copy
已复制链接
快去分享给好友吧!
我知道了
右上角分享
点击右上角分享
0
联系我们:info@booksci.cn Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。 Copyright © 2023 布克学术 All rights reserved.
京ICP备2023020795号-1
ghs 京公网安备 11010802042870号
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术官方微信