Arpita Ghosh, Katrina Ligett, Aaron Roth, G. Schoenebeck
{"title":"Buying private data without verification","authors":"Arpita Ghosh, Katrina Ligett, Aaron Roth, G. Schoenebeck","doi":"10.1145/2600057.2602902","DOIUrl":null,"url":null,"abstract":"We consider the problem of designing a survey to aggregate non-verifiable information from a privacy-sensitive population: an analyst wants to compute some aggregate statistic from the private bits held by each member of a population, but cannot verify the correctness of the bits reported by participants in his survey. Individuals in the population are strategic agents with a cost for privacy, ie, they not only account for the payments they expect to receive from the mechanism, but also their privacy costs from any information revealed about them by the mechanism's outcome---the computed statistic as well as the payments---to determine their utilities. How can the analyst design payments to obtain an accurate estimate of the population statistic when individuals strategically decide both whether to participate and whether to truthfully report their sensitive information' We design a differentially private peer-prediction mechanism [Miller et al. 2005] that supports accurate estimation of the population statistic as a Bayes-Nash equilibrium in settings where agents have explicit preferences for privacy. The mechanism requires knowledge of the marginal prior distribution on bits bi, but does not need full knowledge of the marginal distribution on the costs ci, instead requiring only an approximate upper bound. Our mechanism guarantees ε-differential privacy to each agent i against any adversary who can observe the statistical estimate output by the mechanism, as well as the payments made to the n-1 other agents j ≠ i. Finally, we show that with slightly more structured assumptions on the privacy cost functions of each agent [Chen et al. 2013], the cost of running the survey goes to 0 as the number of agents diverges.","PeriodicalId":203155,"journal":{"name":"Proceedings of the fifteenth ACM conference on Economics and computation","volume":"14 4 1","pages":"0"},"PeriodicalIF":0.0000,"publicationDate":"2014-04-23","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"62","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"Proceedings of the fifteenth ACM conference on Economics and computation","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.1145/2600057.2602902","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
引用次数: 62
Abstract
We consider the problem of designing a survey to aggregate non-verifiable information from a privacy-sensitive population: an analyst wants to compute some aggregate statistic from the private bits held by each member of a population, but cannot verify the correctness of the bits reported by participants in his survey. Individuals in the population are strategic agents with a cost for privacy, ie, they not only account for the payments they expect to receive from the mechanism, but also their privacy costs from any information revealed about them by the mechanism's outcome---the computed statistic as well as the payments---to determine their utilities. How can the analyst design payments to obtain an accurate estimate of the population statistic when individuals strategically decide both whether to participate and whether to truthfully report their sensitive information' We design a differentially private peer-prediction mechanism [Miller et al. 2005] that supports accurate estimation of the population statistic as a Bayes-Nash equilibrium in settings where agents have explicit preferences for privacy. The mechanism requires knowledge of the marginal prior distribution on bits bi, but does not need full knowledge of the marginal distribution on the costs ci, instead requiring only an approximate upper bound. Our mechanism guarantees ε-differential privacy to each agent i against any adversary who can observe the statistical estimate output by the mechanism, as well as the payments made to the n-1 other agents j ≠ i. Finally, we show that with slightly more structured assumptions on the privacy cost functions of each agent [Chen et al. 2013], the cost of running the survey goes to 0 as the number of agents diverges.
我们考虑设计一个调查来汇总来自隐私敏感群体的不可验证信息的问题:分析人员希望从群体中每个成员持有的私有比特中计算一些汇总统计数据,但无法验证调查参与者报告的比特的正确性。人口中的个人是具有隐私成本的战略代理人,也就是说,他们不仅要考虑他们期望从机制中获得的支付,还要考虑他们的隐私成本,这些隐私成本来自于机制结果(计算的统计数据和支付)所揭示的关于他们的任何信息,以确定他们的效用。当个体策略性地决定是否参与和是否如实报告其敏感信息时,分析师如何设计支付以获得对人口统计数据的准确估计?我们设计了一种不同的私人同行预测机制[Miller等人,2005],该机制支持在代理人具有明确隐私偏好的设置中作为贝叶斯-纳什均衡对人口统计数据的准确估计。该机制需要知道比特bi上的边际先验分布,但不需要完全知道成本ci上的边际分布,而只需要一个近似的上界。我们的机制保证每个代理i的ε-差分隐私不受任何可以观察到该机制统计估计输出的对手的影响,以及对n-1个其他代理j≠i的支付。最后,我们表明,对每个代理的隐私成本函数进行稍微结构化的假设[Chen et al. 2013],随着代理数量的分散,运行调查的成本趋于0。