Tian-Yi Zhou, Matthew Lau, Jizhou Chen, Wenke Lee, Xiaoming Huo
{"title":"Optimal Classification-based Anomaly Detection with Neural Networks: Theory and Practice","authors":"Tian-Yi Zhou, Matthew Lau, Jizhou Chen, Wenke Lee, Xiaoming Huo","doi":"arxiv-2409.08521","DOIUrl":null,"url":null,"abstract":"Anomaly detection is an important problem in many application areas, such as\nnetwork security. Many deep learning methods for unsupervised anomaly detection\nproduce good empirical performance but lack theoretical guarantees. By casting\nanomaly detection into a binary classification problem, we establish\nnon-asymptotic upper bounds and a convergence rate on the excess risk on\nrectified linear unit (ReLU) neural networks trained on synthetic anomalies.\nOur convergence rate on the excess risk matches the minimax optimal rate in the\nliterature. Furthermore, we provide lower and upper bounds on the number of\nsynthetic anomalies that can attain this optimality. For practical\nimplementation, we relax some conditions to improve the search for the\nempirical risk minimizer, which leads to competitive performance to other\nclassification-based methods for anomaly detection. Overall, our work provides\nthe first theoretical guarantees of unsupervised neural network-based anomaly\ndetectors and empirical insights on how to design them well.","PeriodicalId":501379,"journal":{"name":"arXiv - STAT - Statistics Theory","volume":"20 1","pages":""},"PeriodicalIF":0.0000,"publicationDate":"2024-09-13","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"arXiv - STAT - Statistics Theory","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/arxiv-2409.08521","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
引用次数: 0
Abstract
Anomaly detection is an important problem in many application areas, such as
network security. Many deep learning methods for unsupervised anomaly detection
produce good empirical performance but lack theoretical guarantees. By casting
anomaly detection into a binary classification problem, we establish
non-asymptotic upper bounds and a convergence rate on the excess risk on
rectified linear unit (ReLU) neural networks trained on synthetic anomalies.
Our convergence rate on the excess risk matches the minimax optimal rate in the
literature. Furthermore, we provide lower and upper bounds on the number of
synthetic anomalies that can attain this optimality. For practical
implementation, we relax some conditions to improve the search for the
empirical risk minimizer, which leads to competitive performance to other
classification-based methods for anomaly detection. Overall, our work provides
the first theoretical guarantees of unsupervised neural network-based anomaly
detectors and empirical insights on how to design them well.