Antonino Ferraro , Gian Marco Orlando , Diego Russo
{"title":"Generative Agent-Based Modeling with Large Language Models for insider threat detection","authors":"Antonino Ferraro , Gian Marco Orlando , Diego Russo","doi":"10.1016/j.engappai.2025.111343","DOIUrl":null,"url":null,"abstract":"<div><div>Insider threats pose a critical challenge in cybersecurity, as individuals within organizations misuse legitimate access to compromise sensitive systems and data. Traditional detection methods often struggle with the complexity of such threats, while Deep Learning (DL) approaches face issues like overfitting and lack of interpretability. To address these limitations, we propose a Generative Agent-Based Modeling (GABM) framework that integrates Large Language Models (LLMs) with a hierarchical multi-agent system.</div><div>Our framework employs Specialized Agents to process categorized log files and generate detailed reports, which are synthesized by a Supervisor Agent for final activity classification. We validated this approach on both network-centric (PicoDomain) and behavior-rich (CERT r5.2) datasets, demonstrating its ability to handle diverse logs, model complex threats, and generalize across insider risk scenarios.</div><div>The framework outperformed existing baselines, prioritizing high recall to minimize false negatives—crucial in cybersecurity contexts. While precision was comparatively lower, this trade-off supports early threat detection. An ablation study highlighted the importance of the Supervisor Agent, whose removal led to a significant drop in performance and increased false positives.</div><div>These results demonstrate the potential of LLM-powered hierarchical multi-agent frameworks for scalable, interpretable, and reliable insider threat detection. Our contributions include the integration of GABM and LLMs, a hierarchical system for log analysis, and the use of Chain-of-Thought reasoning for enhanced interpretability.</div></div>","PeriodicalId":50523,"journal":{"name":"Engineering Applications of Artificial Intelligence","volume":"157 ","pages":"Article 111343"},"PeriodicalIF":7.5000,"publicationDate":"2025-06-19","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"Engineering Applications of Artificial Intelligence","FirstCategoryId":"94","ListUrlMain":"https://www.sciencedirect.com/science/article/pii/S0952197625013454","RegionNum":2,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"Q1","JCRName":"AUTOMATION & CONTROL SYSTEMS","Score":null,"Total":0}
引用次数: 0
Abstract
Insider threats pose a critical challenge in cybersecurity, as individuals within organizations misuse legitimate access to compromise sensitive systems and data. Traditional detection methods often struggle with the complexity of such threats, while Deep Learning (DL) approaches face issues like overfitting and lack of interpretability. To address these limitations, we propose a Generative Agent-Based Modeling (GABM) framework that integrates Large Language Models (LLMs) with a hierarchical multi-agent system.
Our framework employs Specialized Agents to process categorized log files and generate detailed reports, which are synthesized by a Supervisor Agent for final activity classification. We validated this approach on both network-centric (PicoDomain) and behavior-rich (CERT r5.2) datasets, demonstrating its ability to handle diverse logs, model complex threats, and generalize across insider risk scenarios.
The framework outperformed existing baselines, prioritizing high recall to minimize false negatives—crucial in cybersecurity contexts. While precision was comparatively lower, this trade-off supports early threat detection. An ablation study highlighted the importance of the Supervisor Agent, whose removal led to a significant drop in performance and increased false positives.
These results demonstrate the potential of LLM-powered hierarchical multi-agent frameworks for scalable, interpretable, and reliable insider threat detection. Our contributions include the integration of GABM and LLMs, a hierarchical system for log analysis, and the use of Chain-of-Thought reasoning for enhanced interpretability.
期刊介绍:
Artificial Intelligence (AI) is pivotal in driving the fourth industrial revolution, witnessing remarkable advancements across various machine learning methodologies. AI techniques have become indispensable tools for practicing engineers, enabling them to tackle previously insurmountable challenges. Engineering Applications of Artificial Intelligence serves as a global platform for the swift dissemination of research elucidating the practical application of AI methods across all engineering disciplines. Submitted papers are expected to present novel aspects of AI utilized in real-world engineering applications, validated using publicly available datasets to ensure the replicability of research outcomes. Join us in exploring the transformative potential of AI in engineering.