Gautam Chandrasekaran, Adam Klivans, Vasilis Kontonis, Raghu Meka, Konstantinos Stavropoulos
{"title":"学习低内在维度概念的平滑分析法","authors":"Gautam Chandrasekaran, Adam Klivans, Vasilis Kontonis, Raghu Meka, Konstantinos Stavropoulos","doi":"arxiv-2407.00966","DOIUrl":null,"url":null,"abstract":"In traditional models of supervised learning, the goal of a learner -- given\nexamples from an arbitrary joint distribution on $\\mathbb{R}^d \\times \\{\\pm\n1\\}$ -- is to output a hypothesis that is competitive (to within $\\epsilon$) of\nthe best fitting concept from some class. In order to escape strong hardness\nresults for learning even simple concept classes, we introduce a\nsmoothed-analysis framework that requires a learner to compete only with the\nbest classifier that is robust to small random Gaussian perturbation. This subtle change allows us to give a wide array of learning results for any\nconcept that (1) depends on a low-dimensional subspace (aka multi-index model)\nand (2) has a bounded Gaussian surface area. This class includes functions of\nhalfspaces and (low-dimensional) convex sets, cases that are only known to be\nlearnable in non-smoothed settings with respect to highly structured\ndistributions such as Gaussians. Surprisingly, our analysis also yields new results for traditional\nnon-smoothed frameworks such as learning with margin. In particular, we obtain\nthe first algorithm for agnostically learning intersections of $k$-halfspaces\nin time $k^{poly(\\frac{\\log k}{\\epsilon \\gamma}) }$ where $\\gamma$ is the\nmargin parameter. Before our work, the best-known runtime was exponential in\n$k$ (Arriaga and Vempala, 1999).","PeriodicalId":501024,"journal":{"name":"arXiv - CS - Computational Complexity","volume":"18 1","pages":""},"PeriodicalIF":0.0000,"publicationDate":"2024-07-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":"{\"title\":\"Smoothed Analysis for Learning Concepts with Low Intrinsic Dimension\",\"authors\":\"Gautam Chandrasekaran, Adam Klivans, Vasilis Kontonis, Raghu Meka, Konstantinos Stavropoulos\",\"doi\":\"arxiv-2407.00966\",\"DOIUrl\":null,\"url\":null,\"abstract\":\"In traditional models of supervised learning, the goal of a learner -- given\\nexamples from an arbitrary joint distribution on $\\\\mathbb{R}^d \\\\times \\\\{\\\\pm\\n1\\\\}$ -- is to output a hypothesis that is competitive (to within $\\\\epsilon$) of\\nthe best fitting concept from some class. In order to escape strong hardness\\nresults for learning even simple concept classes, we introduce a\\nsmoothed-analysis framework that requires a learner to compete only with the\\nbest classifier that is robust to small random Gaussian perturbation. This subtle change allows us to give a wide array of learning results for any\\nconcept that (1) depends on a low-dimensional subspace (aka multi-index model)\\nand (2) has a bounded Gaussian surface area. This class includes functions of\\nhalfspaces and (low-dimensional) convex sets, cases that are only known to be\\nlearnable in non-smoothed settings with respect to highly structured\\ndistributions such as Gaussians. Surprisingly, our analysis also yields new results for traditional\\nnon-smoothed frameworks such as learning with margin. In particular, we obtain\\nthe first algorithm for agnostically learning intersections of $k$-halfspaces\\nin time $k^{poly(\\\\frac{\\\\log k}{\\\\epsilon \\\\gamma}) }$ where $\\\\gamma$ is the\\nmargin parameter. Before our work, the best-known runtime was exponential in\\n$k$ (Arriaga and Vempala, 1999).\",\"PeriodicalId\":501024,\"journal\":{\"name\":\"arXiv - CS - Computational Complexity\",\"volume\":\"18 1\",\"pages\":\"\"},\"PeriodicalIF\":0.0000,\"publicationDate\":\"2024-07-01\",\"publicationTypes\":\"Journal Article\",\"fieldsOfStudy\":null,\"isOpenAccess\":false,\"openAccessPdf\":\"\",\"citationCount\":\"0\",\"resultStr\":null,\"platform\":\"Semanticscholar\",\"paperid\":null,\"PeriodicalName\":\"arXiv - CS - Computational Complexity\",\"FirstCategoryId\":\"1085\",\"ListUrlMain\":\"https://doi.org/arxiv-2407.00966\",\"RegionNum\":0,\"RegionCategory\":null,\"ArticlePicture\":[],\"TitleCN\":null,\"AbstractTextCN\":null,\"PMCID\":null,\"EPubDate\":\"\",\"PubModel\":\"\",\"JCR\":\"\",\"JCRName\":\"\",\"Score\":null,\"Total\":0}","platform":"Semanticscholar","paperid":null,"PeriodicalName":"arXiv - CS - Computational Complexity","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/arxiv-2407.00966","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
Smoothed Analysis for Learning Concepts with Low Intrinsic Dimension
In traditional models of supervised learning, the goal of a learner -- given
examples from an arbitrary joint distribution on $\mathbb{R}^d \times \{\pm
1\}$ -- is to output a hypothesis that is competitive (to within $\epsilon$) of
the best fitting concept from some class. In order to escape strong hardness
results for learning even simple concept classes, we introduce a
smoothed-analysis framework that requires a learner to compete only with the
best classifier that is robust to small random Gaussian perturbation. This subtle change allows us to give a wide array of learning results for any
concept that (1) depends on a low-dimensional subspace (aka multi-index model)
and (2) has a bounded Gaussian surface area. This class includes functions of
halfspaces and (low-dimensional) convex sets, cases that are only known to be
learnable in non-smoothed settings with respect to highly structured
distributions such as Gaussians. Surprisingly, our analysis also yields new results for traditional
non-smoothed frameworks such as learning with margin. In particular, we obtain
the first algorithm for agnostically learning intersections of $k$-halfspaces
in time $k^{poly(\frac{\log k}{\epsilon \gamma}) }$ where $\gamma$ is the
margin parameter. Before our work, the best-known runtime was exponential in
$k$ (Arriaga and Vempala, 1999).