{"title":"通过投影私人学习超立方体上的平滑分布","authors":"Clément LalanneTSE-R, Sébastien GadatTSE-R, IUF","doi":"arxiv-2409.10083","DOIUrl":null,"url":null,"abstract":"Fueled by the ever-increasing need for statistics that guarantee the privacy\nof their training sets, this article studies the centrally-private estimation\nof Sobolev-smooth densities of probability over the hypercube in dimension d.\nThe contributions of this article are two-fold : Firstly, it generalizes the\none dimensional results of (Lalanne et al., 2023) to non-integer levels of\nsmoothness and to a high-dimensional setting, which is important for two\nreasons : it is more suited for modern learning tasks, and it allows\nunderstanding the relations between privacy, dimensionality and smoothness,\nwhich is a central question with differential privacy. Secondly, this article\npresents a private strategy of estimation that is data-driven (usually referred\nto as adaptive in Statistics) in order to privately choose an estimator that\nachieves a good bias-variance trade-off among a finite family of private\nprojection estimators without prior knowledge of the ground-truth smoothness\n$\\beta$. This is achieved by adapting the Lepskii method for private selection,\nby adding a new penalization term that makes the estimation privacy-aware.","PeriodicalId":501379,"journal":{"name":"arXiv - STAT - Statistics Theory","volume":"39 1","pages":""},"PeriodicalIF":0.0000,"publicationDate":"2024-09-16","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":"{\"title\":\"Privately Learning Smooth Distributions on the Hypercube by Projections\",\"authors\":\"Clément LalanneTSE-R, Sébastien GadatTSE-R, IUF\",\"doi\":\"arxiv-2409.10083\",\"DOIUrl\":null,\"url\":null,\"abstract\":\"Fueled by the ever-increasing need for statistics that guarantee the privacy\\nof their training sets, this article studies the centrally-private estimation\\nof Sobolev-smooth densities of probability over the hypercube in dimension d.\\nThe contributions of this article are two-fold : Firstly, it generalizes the\\none dimensional results of (Lalanne et al., 2023) to non-integer levels of\\nsmoothness and to a high-dimensional setting, which is important for two\\nreasons : it is more suited for modern learning tasks, and it allows\\nunderstanding the relations between privacy, dimensionality and smoothness,\\nwhich is a central question with differential privacy. Secondly, this article\\npresents a private strategy of estimation that is data-driven (usually referred\\nto as adaptive in Statistics) in order to privately choose an estimator that\\nachieves a good bias-variance trade-off among a finite family of private\\nprojection estimators without prior knowledge of the ground-truth smoothness\\n$\\\\beta$. This is achieved by adapting the Lepskii method for private selection,\\nby adding a new penalization term that makes the estimation privacy-aware.\",\"PeriodicalId\":501379,\"journal\":{\"name\":\"arXiv - STAT - Statistics Theory\",\"volume\":\"39 1\",\"pages\":\"\"},\"PeriodicalIF\":0.0000,\"publicationDate\":\"2024-09-16\",\"publicationTypes\":\"Journal Article\",\"fieldsOfStudy\":null,\"isOpenAccess\":false,\"openAccessPdf\":\"\",\"citationCount\":\"0\",\"resultStr\":null,\"platform\":\"Semanticscholar\",\"paperid\":null,\"PeriodicalName\":\"arXiv - STAT - Statistics Theory\",\"FirstCategoryId\":\"1085\",\"ListUrlMain\":\"https://doi.org/arxiv-2409.10083\",\"RegionNum\":0,\"RegionCategory\":null,\"ArticlePicture\":[],\"TitleCN\":null,\"AbstractTextCN\":null,\"PMCID\":null,\"EPubDate\":\"\",\"PubModel\":\"\",\"JCR\":\"\",\"JCRName\":\"\",\"Score\":null,\"Total\":0}","platform":"Semanticscholar","paperid":null,"PeriodicalName":"arXiv - STAT - Statistics Theory","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/arxiv-2409.10083","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
引用次数: 0
摘要
随着人们对保证训练集隐私的统计的需求日益增长,本文研究了在维数为 d 的超立方体上对 Sobolev 平滑概率密度的集中隐私估计、首先,本文将(Lalanne 等人,2023 年)的一维结果推广到非整数平滑度水平和高维环境,这有两个重要原因:一是它更适合现代学习任务,二是它允许理解隐私、维度和平滑度之间的关系,而这是微分隐私的核心问题。其次,本文介绍了一种由数据驱动的私人估计策略(通常在统计学中称为自适应策略),以便在事先不知道地面真实平滑度$beta$的情况下,在有限的私人投影估计器家族中私下选择一个能实现良好偏差-方差权衡的估计器。这是通过调整用于私人选择的 Lepskii 方法来实现的,方法是添加一个新的惩罚项,使估计具有隐私意识。
Privately Learning Smooth Distributions on the Hypercube by Projections
Fueled by the ever-increasing need for statistics that guarantee the privacy
of their training sets, this article studies the centrally-private estimation
of Sobolev-smooth densities of probability over the hypercube in dimension d.
The contributions of this article are two-fold : Firstly, it generalizes the
one dimensional results of (Lalanne et al., 2023) to non-integer levels of
smoothness and to a high-dimensional setting, which is important for two
reasons : it is more suited for modern learning tasks, and it allows
understanding the relations between privacy, dimensionality and smoothness,
which is a central question with differential privacy. Secondly, this article
presents a private strategy of estimation that is data-driven (usually referred
to as adaptive in Statistics) in order to privately choose an estimator that
achieves a good bias-variance trade-off among a finite family of private
projection estimators without prior knowledge of the ground-truth smoothness
$\beta$. This is achieved by adapting the Lepskii method for private selection,
by adding a new penalization term that makes the estimation privacy-aware.