Guy Blanc, Alexandre Hayderi, Caleb Koch, Li-Yang Tan
{"title":"平滑提升的采样复杂性与硬核定理的严密性","authors":"Guy Blanc, Alexandre Hayderi, Caleb Koch, Li-Yang Tan","doi":"arxiv-2409.11597","DOIUrl":null,"url":null,"abstract":"Smooth boosters generate distributions that do not place too much weight on\nany given example. Originally introduced for their noise-tolerant properties,\nsuch boosters have also found applications in differential privacy,\nreproducibility, and quantum learning theory. We study and settle the sample\ncomplexity of smooth boosting: we exhibit a class that can be weak learned to\n$\\gamma$-advantage over smooth distributions with $m$ samples, for which strong\nlearning over the uniform distribution requires\n$\\tilde{\\Omega}(1/\\gamma^2)\\cdot m$ samples. This matches the overhead of\nexisting smooth boosters and provides the first separation from the setting of\ndistribution-independent boosting, for which the corresponding overhead is\n$O(1/\\gamma)$. Our work also sheds new light on Impagliazzo's hardcore theorem from\ncomplexity theory, all known proofs of which can be cast in the framework of\nsmooth boosting. For a function $f$ that is mildly hard against size-$s$\ncircuits, the hardcore theorem provides a set of inputs on which $f$ is\nextremely hard against size-$s'$ circuits. A downside of this important result\nis the loss in circuit size, i.e. that $s' \\ll s$. Answering a question of\nTrevisan, we show that this size loss is necessary and in fact, the parameters\nachieved by known proofs are the best possible.","PeriodicalId":501340,"journal":{"name":"arXiv - STAT - Machine Learning","volume":"74 1","pages":""},"PeriodicalIF":0.0000,"publicationDate":"2024-09-17","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":"{\"title\":\"The Sample Complexity of Smooth Boosting and the Tightness of the Hardcore Theorem\",\"authors\":\"Guy Blanc, Alexandre Hayderi, Caleb Koch, Li-Yang Tan\",\"doi\":\"arxiv-2409.11597\",\"DOIUrl\":null,\"url\":null,\"abstract\":\"Smooth boosters generate distributions that do not place too much weight on\\nany given example. Originally introduced for their noise-tolerant properties,\\nsuch boosters have also found applications in differential privacy,\\nreproducibility, and quantum learning theory. We study and settle the sample\\ncomplexity of smooth boosting: we exhibit a class that can be weak learned to\\n$\\\\gamma$-advantage over smooth distributions with $m$ samples, for which strong\\nlearning over the uniform distribution requires\\n$\\\\tilde{\\\\Omega}(1/\\\\gamma^2)\\\\cdot m$ samples. This matches the overhead of\\nexisting smooth boosters and provides the first separation from the setting of\\ndistribution-independent boosting, for which the corresponding overhead is\\n$O(1/\\\\gamma)$. Our work also sheds new light on Impagliazzo's hardcore theorem from\\ncomplexity theory, all known proofs of which can be cast in the framework of\\nsmooth boosting. For a function $f$ that is mildly hard against size-$s$\\ncircuits, the hardcore theorem provides a set of inputs on which $f$ is\\nextremely hard against size-$s'$ circuits. A downside of this important result\\nis the loss in circuit size, i.e. that $s' \\\\ll s$. Answering a question of\\nTrevisan, we show that this size loss is necessary and in fact, the parameters\\nachieved by known proofs are the best possible.\",\"PeriodicalId\":501340,\"journal\":{\"name\":\"arXiv - STAT - Machine Learning\",\"volume\":\"74 1\",\"pages\":\"\"},\"PeriodicalIF\":0.0000,\"publicationDate\":\"2024-09-17\",\"publicationTypes\":\"Journal Article\",\"fieldsOfStudy\":null,\"isOpenAccess\":false,\"openAccessPdf\":\"\",\"citationCount\":\"0\",\"resultStr\":null,\"platform\":\"Semanticscholar\",\"paperid\":null,\"PeriodicalName\":\"arXiv - STAT - Machine Learning\",\"FirstCategoryId\":\"1085\",\"ListUrlMain\":\"https://doi.org/arxiv-2409.11597\",\"RegionNum\":0,\"RegionCategory\":null,\"ArticlePicture\":[],\"TitleCN\":null,\"AbstractTextCN\":null,\"PMCID\":null,\"EPubDate\":\"\",\"PubModel\":\"\",\"JCR\":\"\",\"JCRName\":\"\",\"Score\":null,\"Total\":0}","platform":"Semanticscholar","paperid":null,"PeriodicalName":"arXiv - STAT - Machine Learning","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/arxiv-2409.11597","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
The Sample Complexity of Smooth Boosting and the Tightness of the Hardcore Theorem
Smooth boosters generate distributions that do not place too much weight on
any given example. Originally introduced for their noise-tolerant properties,
such boosters have also found applications in differential privacy,
reproducibility, and quantum learning theory. We study and settle the sample
complexity of smooth boosting: we exhibit a class that can be weak learned to
$\gamma$-advantage over smooth distributions with $m$ samples, for which strong
learning over the uniform distribution requires
$\tilde{\Omega}(1/\gamma^2)\cdot m$ samples. This matches the overhead of
existing smooth boosters and provides the first separation from the setting of
distribution-independent boosting, for which the corresponding overhead is
$O(1/\gamma)$. Our work also sheds new light on Impagliazzo's hardcore theorem from
complexity theory, all known proofs of which can be cast in the framework of
smooth boosting. For a function $f$ that is mildly hard against size-$s$
circuits, the hardcore theorem provides a set of inputs on which $f$ is
extremely hard against size-$s'$ circuits. A downside of this important result
is the loss in circuit size, i.e. that $s' \ll s$. Answering a question of
Trevisan, we show that this size loss is necessary and in fact, the parameters
achieved by known proofs are the best possible.