{"title":"Statistically Valid Information Bottleneck via Multiple Hypothesis Testing","authors":"Amirmohammad Farzaneh, Osvaldo Simeone","doi":"arxiv-2409.07325","DOIUrl":null,"url":null,"abstract":"The information bottleneck (IB) problem is a widely studied framework in\nmachine learning for extracting compressed features that are informative for\ndownstream tasks. However, current approaches to solving the IB problem rely on\na heuristic tuning of hyperparameters, offering no guarantees that the learned\nfeatures satisfy information-theoretic constraints. In this work, we introduce\na statistically valid solution to this problem, referred to as IB via multiple\nhypothesis testing (IB-MHT), which ensures that the learned features meet the\nIB constraints with high probability, regardless of the size of the available\ndataset. The proposed methodology builds on Pareto testing and learn-then-test\n(LTT), and it wraps around existing IB solvers to provide statistical\nguarantees on the IB constraints. We demonstrate the performance of IB-MHT on\nclassical and deterministic IB formulations, validating the effectiveness of\nIB-MHT in outperforming conventional methods in terms of statistical robustness\nand reliability.","PeriodicalId":501082,"journal":{"name":"arXiv - MATH - Information Theory","volume":"4 1","pages":""},"PeriodicalIF":0.0000,"publicationDate":"2024-09-11","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"arXiv - MATH - Information Theory","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/arxiv-2409.07325","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
引用次数: 0
Abstract
The information bottleneck (IB) problem is a widely studied framework in
machine learning for extracting compressed features that are informative for
downstream tasks. However, current approaches to solving the IB problem rely on
a heuristic tuning of hyperparameters, offering no guarantees that the learned
features satisfy information-theoretic constraints. In this work, we introduce
a statistically valid solution to this problem, referred to as IB via multiple
hypothesis testing (IB-MHT), which ensures that the learned features meet the
IB constraints with high probability, regardless of the size of the available
dataset. The proposed methodology builds on Pareto testing and learn-then-test
(LTT), and it wraps around existing IB solvers to provide statistical
guarantees on the IB constraints. We demonstrate the performance of IB-MHT on
classical and deterministic IB formulations, validating the effectiveness of
IB-MHT in outperforming conventional methods in terms of statistical robustness
and reliability.