Guy Blanc, Alexandre Hayderi, Caleb Koch, Li-Yang Tan
{"title":"The Sample Complexity of Smooth Boosting and the Tightness of the Hardcore Theorem","authors":"Guy Blanc, Alexandre Hayderi, Caleb Koch, Li-Yang Tan","doi":"arxiv-2409.11597","DOIUrl":null,"url":null,"abstract":"Smooth boosters generate distributions that do not place too much weight on\nany given example. Originally introduced for their noise-tolerant properties,\nsuch boosters have also found applications in differential privacy,\nreproducibility, and quantum learning theory. We study and settle the sample\ncomplexity of smooth boosting: we exhibit a class that can be weak learned to\n$\\gamma$-advantage over smooth distributions with $m$ samples, for which strong\nlearning over the uniform distribution requires\n$\\tilde{\\Omega}(1/\\gamma^2)\\cdot m$ samples. This matches the overhead of\nexisting smooth boosters and provides the first separation from the setting of\ndistribution-independent boosting, for which the corresponding overhead is\n$O(1/\\gamma)$. Our work also sheds new light on Impagliazzo's hardcore theorem from\ncomplexity theory, all known proofs of which can be cast in the framework of\nsmooth boosting. For a function $f$ that is mildly hard against size-$s$\ncircuits, the hardcore theorem provides a set of inputs on which $f$ is\nextremely hard against size-$s'$ circuits. A downside of this important result\nis the loss in circuit size, i.e. that $s' \\ll s$. Answering a question of\nTrevisan, we show that this size loss is necessary and in fact, the parameters\nachieved by known proofs are the best possible.","PeriodicalId":501340,"journal":{"name":"arXiv - STAT - Machine Learning","volume":"74 1","pages":""},"PeriodicalIF":0.0000,"publicationDate":"2024-09-17","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"arXiv - STAT - Machine Learning","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/arxiv-2409.11597","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
引用次数: 0
Abstract
Smooth boosters generate distributions that do not place too much weight on
any given example. Originally introduced for their noise-tolerant properties,
such boosters have also found applications in differential privacy,
reproducibility, and quantum learning theory. We study and settle the sample
complexity of smooth boosting: we exhibit a class that can be weak learned to
$\gamma$-advantage over smooth distributions with $m$ samples, for which strong
learning over the uniform distribution requires
$\tilde{\Omega}(1/\gamma^2)\cdot m$ samples. This matches the overhead of
existing smooth boosters and provides the first separation from the setting of
distribution-independent boosting, for which the corresponding overhead is
$O(1/\gamma)$. Our work also sheds new light on Impagliazzo's hardcore theorem from
complexity theory, all known proofs of which can be cast in the framework of
smooth boosting. For a function $f$ that is mildly hard against size-$s$
circuits, the hardcore theorem provides a set of inputs on which $f$ is
extremely hard against size-$s'$ circuits. A downside of this important result
is the loss in circuit size, i.e. that $s' \ll s$. Answering a question of
Trevisan, we show that this size loss is necessary and in fact, the parameters
achieved by known proofs are the best possible.