{"title":"分数奈维贝叶 (FNB):针对简明加权选择性奈维贝叶分类器的非凸优化","authors":"Carine Hue, Marc Boullé","doi":"arxiv-2409.11100","DOIUrl":null,"url":null,"abstract":"We study supervised classification for datasets with a very large number of\ninput variables. The na\\\"ive Bayes classifier is attractive for its simplicity,\nscalability and effectiveness in many real data applications. When the strong\nna\\\"ive Bayes assumption of conditional independence of the input variables\ngiven the target variable is not valid, variable selection and model averaging\nare two common ways to improve the performance. In the case of the na\\\"ive\nBayes classifier, the resulting weighting scheme on the models reduces to a\nweighting scheme on the variables. Here we focus on direct estimation of\nvariable weights in such a weighted na\\\"ive Bayes classifier. We propose a\nsparse regularization of the model log-likelihood, which takes into account\nprior penalization costs related to each input variable. Compared to averaging\nbased classifiers used up until now, our main goal is to obtain parsimonious\nrobust models with less variables and equivalent performance. The direct\nestimation of the variable weights amounts to a non-convex optimization problem\nfor which we propose and compare several two-stage algorithms. First, the\ncriterion obtained by convex relaxation is minimized using several variants of\nstandard gradient methods. Then, the initial non-convex optimization problem is\nsolved using local optimization methods initialized with the result of the\nfirst stage. The various proposed algorithms result in optimization-based\nweighted na\\\"ive Bayes classifiers, that are evaluated on benchmark datasets\nand positioned w.r.t. to a reference averaging-based classifier.","PeriodicalId":501340,"journal":{"name":"arXiv - STAT - Machine Learning","volume":"94 1","pages":""},"PeriodicalIF":0.0000,"publicationDate":"2024-09-17","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":"{\"title\":\"Fractional Naive Bayes (FNB): non-convex optimization for a parsimonious weighted selective naive Bayes classifier\",\"authors\":\"Carine Hue, Marc Boullé\",\"doi\":\"arxiv-2409.11100\",\"DOIUrl\":null,\"url\":null,\"abstract\":\"We study supervised classification for datasets with a very large number of\\ninput variables. The na\\\\\\\"ive Bayes classifier is attractive for its simplicity,\\nscalability and effectiveness in many real data applications. When the strong\\nna\\\\\\\"ive Bayes assumption of conditional independence of the input variables\\ngiven the target variable is not valid, variable selection and model averaging\\nare two common ways to improve the performance. In the case of the na\\\\\\\"ive\\nBayes classifier, the resulting weighting scheme on the models reduces to a\\nweighting scheme on the variables. Here we focus on direct estimation of\\nvariable weights in such a weighted na\\\\\\\"ive Bayes classifier. We propose a\\nsparse regularization of the model log-likelihood, which takes into account\\nprior penalization costs related to each input variable. Compared to averaging\\nbased classifiers used up until now, our main goal is to obtain parsimonious\\nrobust models with less variables and equivalent performance. The direct\\nestimation of the variable weights amounts to a non-convex optimization problem\\nfor which we propose and compare several two-stage algorithms. First, the\\ncriterion obtained by convex relaxation is minimized using several variants of\\nstandard gradient methods. Then, the initial non-convex optimization problem is\\nsolved using local optimization methods initialized with the result of the\\nfirst stage. The various proposed algorithms result in optimization-based\\nweighted na\\\\\\\"ive Bayes classifiers, that are evaluated on benchmark datasets\\nand positioned w.r.t. to a reference averaging-based classifier.\",\"PeriodicalId\":501340,\"journal\":{\"name\":\"arXiv - STAT - Machine Learning\",\"volume\":\"94 1\",\"pages\":\"\"},\"PeriodicalIF\":0.0000,\"publicationDate\":\"2024-09-17\",\"publicationTypes\":\"Journal Article\",\"fieldsOfStudy\":null,\"isOpenAccess\":false,\"openAccessPdf\":\"\",\"citationCount\":\"0\",\"resultStr\":null,\"platform\":\"Semanticscholar\",\"paperid\":null,\"PeriodicalName\":\"arXiv - STAT - Machine Learning\",\"FirstCategoryId\":\"1085\",\"ListUrlMain\":\"https://doi.org/arxiv-2409.11100\",\"RegionNum\":0,\"RegionCategory\":null,\"ArticlePicture\":[],\"TitleCN\":null,\"AbstractTextCN\":null,\"PMCID\":null,\"EPubDate\":\"\",\"PubModel\":\"\",\"JCR\":\"\",\"JCRName\":\"\",\"Score\":null,\"Total\":0}","platform":"Semanticscholar","paperid":null,"PeriodicalName":"arXiv - STAT - Machine Learning","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/arxiv-2409.11100","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
Fractional Naive Bayes (FNB): non-convex optimization for a parsimonious weighted selective naive Bayes classifier
We study supervised classification for datasets with a very large number of
input variables. The na\"ive Bayes classifier is attractive for its simplicity,
scalability and effectiveness in many real data applications. When the strong
na\"ive Bayes assumption of conditional independence of the input variables
given the target variable is not valid, variable selection and model averaging
are two common ways to improve the performance. In the case of the na\"ive
Bayes classifier, the resulting weighting scheme on the models reduces to a
weighting scheme on the variables. Here we focus on direct estimation of
variable weights in such a weighted na\"ive Bayes classifier. We propose a
sparse regularization of the model log-likelihood, which takes into account
prior penalization costs related to each input variable. Compared to averaging
based classifiers used up until now, our main goal is to obtain parsimonious
robust models with less variables and equivalent performance. The direct
estimation of the variable weights amounts to a non-convex optimization problem
for which we propose and compare several two-stage algorithms. First, the
criterion obtained by convex relaxation is minimized using several variants of
standard gradient methods. Then, the initial non-convex optimization problem is
solved using local optimization methods initialized with the result of the
first stage. The various proposed algorithms result in optimization-based
weighted na\"ive Bayes classifiers, that are evaluated on benchmark datasets
and positioned w.r.t. to a reference averaging-based classifier.