在线分类中的大数对抗律与最优后悔

IF 1.2 3区 计算机科学 Q3 COMPUTER SCIENCE, THEORY & METHODS
Noga Alon, Omri Ben-Eliezer, Yuval Dagan, Shay Moran, Moni Naor, Eylon Yogev
{"title":"在线分类中的大数对抗律与最优后悔","authors":"Noga Alon, Omri Ben-Eliezer, Yuval Dagan, Shay Moran, Moni Naor, Eylon Yogev","doi":"10.1137/21m1441924","DOIUrl":null,"url":null,"abstract":"Laws of large numbers guarantee that given a large enough sample from some population, the measure of any fixed subpopulation is well-estimated by its frequency in the sample. We study laws of large numbers in sampling processes that can affect the environment they are acting upon and interact with it. Specifically, we consider the sequential sampling model proposed by [O. Ben-Eliezer and E. Yogev, The adversarial robustness of sampling, in Proceedings of the 39th ACM SIGMOD-SIGACT-SIGAI Symposium on Principles of Database Systems (PODS), 2020, pp. 49–62] and characterize the classes which admit a uniform law of large numbers in this model: these are exactly the classes that are online learnable. Our characterization may be interpreted as an online analogue to the equivalence between learnability and uniform convergence in statistical (PAC) learning. The sample-complexity bounds we obtain are tight for many parameter regimes, and as an application, we determine the optimal regret bounds in online learning, stated in terms of Littlestone’s dimension, thus resolving the main open question from [S. Ben-David, D. Pál, and S. Shalev-Shwartz, Agnostic online learning, in Proceedings of the 22nd Conference on Learning Theory (COLT), 2009], which was also posed by [A. Rakhlin, K. Sridharan, and A. Tewari, J. Mach. Learn. Res., 16 (2015), pp. 155–186].","PeriodicalId":49532,"journal":{"name":"SIAM Journal on Computing","volume":"243 7","pages":"0"},"PeriodicalIF":1.2000,"publicationDate":"2023-11-03","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"2","resultStr":"{\"title\":\"Adversarial Laws of Large Numbers and Optimal Regret in Online Classification\",\"authors\":\"Noga Alon, Omri Ben-Eliezer, Yuval Dagan, Shay Moran, Moni Naor, Eylon Yogev\",\"doi\":\"10.1137/21m1441924\",\"DOIUrl\":null,\"url\":null,\"abstract\":\"Laws of large numbers guarantee that given a large enough sample from some population, the measure of any fixed subpopulation is well-estimated by its frequency in the sample. We study laws of large numbers in sampling processes that can affect the environment they are acting upon and interact with it. Specifically, we consider the sequential sampling model proposed by [O. Ben-Eliezer and E. Yogev, The adversarial robustness of sampling, in Proceedings of the 39th ACM SIGMOD-SIGACT-SIGAI Symposium on Principles of Database Systems (PODS), 2020, pp. 49–62] and characterize the classes which admit a uniform law of large numbers in this model: these are exactly the classes that are online learnable. Our characterization may be interpreted as an online analogue to the equivalence between learnability and uniform convergence in statistical (PAC) learning. The sample-complexity bounds we obtain are tight for many parameter regimes, and as an application, we determine the optimal regret bounds in online learning, stated in terms of Littlestone’s dimension, thus resolving the main open question from [S. Ben-David, D. Pál, and S. Shalev-Shwartz, Agnostic online learning, in Proceedings of the 22nd Conference on Learning Theory (COLT), 2009], which was also posed by [A. Rakhlin, K. Sridharan, and A. Tewari, J. Mach. Learn. Res., 16 (2015), pp. 155–186].\",\"PeriodicalId\":49532,\"journal\":{\"name\":\"SIAM Journal on Computing\",\"volume\":\"243 7\",\"pages\":\"0\"},\"PeriodicalIF\":1.2000,\"publicationDate\":\"2023-11-03\",\"publicationTypes\":\"Journal Article\",\"fieldsOfStudy\":null,\"isOpenAccess\":false,\"openAccessPdf\":\"\",\"citationCount\":\"2\",\"resultStr\":null,\"platform\":\"Semanticscholar\",\"paperid\":null,\"PeriodicalName\":\"SIAM Journal on Computing\",\"FirstCategoryId\":\"1085\",\"ListUrlMain\":\"https://doi.org/10.1137/21m1441924\",\"RegionNum\":3,\"RegionCategory\":\"计算机科学\",\"ArticlePicture\":[],\"TitleCN\":null,\"AbstractTextCN\":null,\"PMCID\":null,\"EPubDate\":\"\",\"PubModel\":\"\",\"JCR\":\"Q3\",\"JCRName\":\"COMPUTER SCIENCE, THEORY & METHODS\",\"Score\":null,\"Total\":0}","platform":"Semanticscholar","paperid":null,"PeriodicalName":"SIAM Journal on Computing","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.1137/21m1441924","RegionNum":3,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"Q3","JCRName":"COMPUTER SCIENCE, THEORY & METHODS","Score":null,"Total":0}
引用次数: 2

摘要

大数定律保证,给定来自某个群体的足够大的样本,任何固定子群体的测量都可以通过样本中的频率得到很好的估计。我们研究采样过程中的大数定律,这些定律可以影响它们所作用的环境并与之相互作用。具体来说,我们考虑了由[O。Ben-Eliezer和E. Yogev,《抽样的对抗性稳稳性》,发表于第39届ACM SIGMOD-SIGACT-SIGAI数据库系统原理研讨会(PODS), 2020年,第49-62页),并描述了在该模型中承认大数统一定律的类:这些正是在线可学习的类。我们的表征可以被解释为统计(PAC)学习中可学习性和均匀收敛之间等价的在线模拟。我们获得的样本复杂度界限对于许多参数体系都是严格的,并且作为一个应用,我们确定了在线学习中的最优后悔界限,用Littlestone维表示,从而解决了[S]中的主要开放问题。Ben-David, D. Pál, S. shalov - shwartz,不可知论在线学习,第22届学习理论会议论文集(COLT), 2009)。Rakhlin, K. Sridharan和A. Tewari, J. Mach。学习。Res., 16 (2015), pp. 155-186]。
本文章由计算机程序翻译,如有差异,请以英文原文为准。
Adversarial Laws of Large Numbers and Optimal Regret in Online Classification
Laws of large numbers guarantee that given a large enough sample from some population, the measure of any fixed subpopulation is well-estimated by its frequency in the sample. We study laws of large numbers in sampling processes that can affect the environment they are acting upon and interact with it. Specifically, we consider the sequential sampling model proposed by [O. Ben-Eliezer and E. Yogev, The adversarial robustness of sampling, in Proceedings of the 39th ACM SIGMOD-SIGACT-SIGAI Symposium on Principles of Database Systems (PODS), 2020, pp. 49–62] and characterize the classes which admit a uniform law of large numbers in this model: these are exactly the classes that are online learnable. Our characterization may be interpreted as an online analogue to the equivalence between learnability and uniform convergence in statistical (PAC) learning. The sample-complexity bounds we obtain are tight for many parameter regimes, and as an application, we determine the optimal regret bounds in online learning, stated in terms of Littlestone’s dimension, thus resolving the main open question from [S. Ben-David, D. Pál, and S. Shalev-Shwartz, Agnostic online learning, in Proceedings of the 22nd Conference on Learning Theory (COLT), 2009], which was also posed by [A. Rakhlin, K. Sridharan, and A. Tewari, J. Mach. Learn. Res., 16 (2015), pp. 155–186].
求助全文
通过发布文献求助,成功后即可免费获取论文全文。 去求助
来源期刊
SIAM Journal on Computing
SIAM Journal on Computing 工程技术-计算机:理论方法
CiteScore
4.60
自引率
0.00%
发文量
68
审稿时长
6-12 weeks
期刊介绍: The SIAM Journal on Computing aims to provide coverage of the most significant work going on in the mathematical and formal aspects of computer science and nonnumerical computing. Submissions must be clearly written and make a significant technical contribution. Topics include but are not limited to analysis and design of algorithms, algorithmic game theory, data structures, computational complexity, computational algebra, computational aspects of combinatorics and graph theory, computational biology, computational geometry, computational robotics, the mathematical aspects of programming languages, artificial intelligence, computational learning, databases, information retrieval, cryptography, networks, distributed computing, parallel algorithms, and computer architecture.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
确定
请完成安全验证×
copy
已复制链接
快去分享给好友吧!
我知道了
右上角分享
点击右上角分享
0
联系我们:info@booksci.cn Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。 Copyright © 2023 布克学术 All rights reserved.
京ICP备2023020795号-1
ghs 京公网安备 11010802042870号
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术官方微信