平滑提升的采样复杂性与硬核定理的严密性

Guy Blanc, Alexandre Hayderi, Caleb Koch, Li-Yang Tan
{"title":"平滑提升的采样复杂性与硬核定理的严密性","authors":"Guy Blanc, Alexandre Hayderi, Caleb Koch, Li-Yang Tan","doi":"arxiv-2409.11597","DOIUrl":null,"url":null,"abstract":"Smooth boosters generate distributions that do not place too much weight on\nany given example. Originally introduced for their noise-tolerant properties,\nsuch boosters have also found applications in differential privacy,\nreproducibility, and quantum learning theory. We study and settle the sample\ncomplexity of smooth boosting: we exhibit a class that can be weak learned to\n$\\gamma$-advantage over smooth distributions with $m$ samples, for which strong\nlearning over the uniform distribution requires\n$\\tilde{\\Omega}(1/\\gamma^2)\\cdot m$ samples. This matches the overhead of\nexisting smooth boosters and provides the first separation from the setting of\ndistribution-independent boosting, for which the corresponding overhead is\n$O(1/\\gamma)$. Our work also sheds new light on Impagliazzo's hardcore theorem from\ncomplexity theory, all known proofs of which can be cast in the framework of\nsmooth boosting. For a function $f$ that is mildly hard against size-$s$\ncircuits, the hardcore theorem provides a set of inputs on which $f$ is\nextremely hard against size-$s'$ circuits. A downside of this important result\nis the loss in circuit size, i.e. that $s' \\ll s$. Answering a question of\nTrevisan, we show that this size loss is necessary and in fact, the parameters\nachieved by known proofs are the best possible.","PeriodicalId":501340,"journal":{"name":"arXiv - STAT - Machine Learning","volume":"74 1","pages":""},"PeriodicalIF":0.0000,"publicationDate":"2024-09-17","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":"{\"title\":\"The Sample Complexity of Smooth Boosting and the Tightness of the Hardcore Theorem\",\"authors\":\"Guy Blanc, Alexandre Hayderi, Caleb Koch, Li-Yang Tan\",\"doi\":\"arxiv-2409.11597\",\"DOIUrl\":null,\"url\":null,\"abstract\":\"Smooth boosters generate distributions that do not place too much weight on\\nany given example. Originally introduced for their noise-tolerant properties,\\nsuch boosters have also found applications in differential privacy,\\nreproducibility, and quantum learning theory. We study and settle the sample\\ncomplexity of smooth boosting: we exhibit a class that can be weak learned to\\n$\\\\gamma$-advantage over smooth distributions with $m$ samples, for which strong\\nlearning over the uniform distribution requires\\n$\\\\tilde{\\\\Omega}(1/\\\\gamma^2)\\\\cdot m$ samples. This matches the overhead of\\nexisting smooth boosters and provides the first separation from the setting of\\ndistribution-independent boosting, for which the corresponding overhead is\\n$O(1/\\\\gamma)$. Our work also sheds new light on Impagliazzo's hardcore theorem from\\ncomplexity theory, all known proofs of which can be cast in the framework of\\nsmooth boosting. For a function $f$ that is mildly hard against size-$s$\\ncircuits, the hardcore theorem provides a set of inputs on which $f$ is\\nextremely hard against size-$s'$ circuits. A downside of this important result\\nis the loss in circuit size, i.e. that $s' \\\\ll s$. Answering a question of\\nTrevisan, we show that this size loss is necessary and in fact, the parameters\\nachieved by known proofs are the best possible.\",\"PeriodicalId\":501340,\"journal\":{\"name\":\"arXiv - STAT - Machine Learning\",\"volume\":\"74 1\",\"pages\":\"\"},\"PeriodicalIF\":0.0000,\"publicationDate\":\"2024-09-17\",\"publicationTypes\":\"Journal Article\",\"fieldsOfStudy\":null,\"isOpenAccess\":false,\"openAccessPdf\":\"\",\"citationCount\":\"0\",\"resultStr\":null,\"platform\":\"Semanticscholar\",\"paperid\":null,\"PeriodicalName\":\"arXiv - STAT - Machine Learning\",\"FirstCategoryId\":\"1085\",\"ListUrlMain\":\"https://doi.org/arxiv-2409.11597\",\"RegionNum\":0,\"RegionCategory\":null,\"ArticlePicture\":[],\"TitleCN\":null,\"AbstractTextCN\":null,\"PMCID\":null,\"EPubDate\":\"\",\"PubModel\":\"\",\"JCR\":\"\",\"JCRName\":\"\",\"Score\":null,\"Total\":0}","platform":"Semanticscholar","paperid":null,"PeriodicalName":"arXiv - STAT - Machine Learning","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/arxiv-2409.11597","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
引用次数: 0

摘要

平滑助推器产生的分布不会对任何给定的例子赋予过多的权重。这类助推器最初是因其噪声容忍特性而被引入的,现在也被应用于差分隐私、可重现性和量子学习理论中。我们研究并解决了平滑提升的样本复杂性问题:我们展示了一类可以在具有 $m$ 样本的平滑分布上弱学习到$\gamma$-advantage,而在均匀分布上强学习需要$\tilde{\Omega}(1/\gamma^2)\cdot m$ 样本。这与现有的平滑助推器的开销相匹配,并首次与独立于分布的助推设置相分离,后者的相应开销为$O(1/\gamma)$。我们的研究还为复杂性理论中的 Impagliazzo 铁杆定理带来了新的启示,所有已知的证明都可以在平滑助推器的框架内进行。对于一个对 size-$s$ 电路有轻微困难的函数 $f$,核心定理提供了一组输入,在这些输入上,$f$ 对 size-$s'$ 电路有极大的困难。这一重要结果的一个缺点是电路规模的损失,即 $s' \ll s$。在回答特雷维桑的一个问题时,我们证明了这种大小损失是必要的,而且事实上,已知证明所达到的参数是可能的最佳参数。
本文章由计算机程序翻译,如有差异,请以英文原文为准。
The Sample Complexity of Smooth Boosting and the Tightness of the Hardcore Theorem
Smooth boosters generate distributions that do not place too much weight on any given example. Originally introduced for their noise-tolerant properties, such boosters have also found applications in differential privacy, reproducibility, and quantum learning theory. We study and settle the sample complexity of smooth boosting: we exhibit a class that can be weak learned to $\gamma$-advantage over smooth distributions with $m$ samples, for which strong learning over the uniform distribution requires $\tilde{\Omega}(1/\gamma^2)\cdot m$ samples. This matches the overhead of existing smooth boosters and provides the first separation from the setting of distribution-independent boosting, for which the corresponding overhead is $O(1/\gamma)$. Our work also sheds new light on Impagliazzo's hardcore theorem from complexity theory, all known proofs of which can be cast in the framework of smooth boosting. For a function $f$ that is mildly hard against size-$s$ circuits, the hardcore theorem provides a set of inputs on which $f$ is extremely hard against size-$s'$ circuits. A downside of this important result is the loss in circuit size, i.e. that $s' \ll s$. Answering a question of Trevisan, we show that this size loss is necessary and in fact, the parameters achieved by known proofs are the best possible.
求助全文
通过发布文献求助,成功后即可免费获取论文全文。 去求助
来源期刊
自引率
0.00%
发文量
0
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
确定
请完成安全验证×
copy
已复制链接
快去分享给好友吧!
我知道了
右上角分享
点击右上角分享
0
联系我们:info@booksci.cn Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。 Copyright © 2023 布克学术 All rights reserved.
京ICP备2023020795号-1
ghs 京公网安备 11010802042870号
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术官方微信