Better lower bounds for locally decodable codes

A. Deshpande, Rahul Jain, T. Kavitha, J. Radhakrishnan, Satyanarayana V. Lokam
{"title":"Better lower bounds for locally decodable codes","authors":"A. Deshpande, Rahul Jain, T. Kavitha, J. Radhakrishnan, Satyanarayana V. Lokam","doi":"10.1109/CCC.2002.1004354","DOIUrl":null,"url":null,"abstract":"An error-correcting code is said to be locally decodable if a randomized algorithm can recover any single bit of a message by reading only a small number of symbols of a possibly corrupted encoding of the message. Katz and Trevisan (2000) showed that any such code C: {0, 1} /spl rarr/ /spl Sigma//sup m/ with a decoding algorithm that makes at most q probes must satisfy m = /spl Omega/((n/log |/spl Sigma/|)/sup q/(q-1)/). They assumed that the decoding algorithm is non-adaptive, and left open the question of proving similar bounds for adaptive decoders. We improve the results of Katz and Trevisan (2000) in two ways. First, we give a more direct proof of their result. Second, and this is our main result, we prove that m = /spl Omega/((n/log|/spl Sigma/|)/sup q/(q-1)/) even if the decoding algorithm is adaptive. An important ingredient of our proof is a randomized method for smoothing an adaptive decoding algorithm. The main technical tool we employ is the Second Moment Method.","PeriodicalId":193513,"journal":{"name":"Proceedings 17th IEEE Annual Conference on Computational Complexity","volume":"49 1","pages":"0"},"PeriodicalIF":0.0000,"publicationDate":"2002-05-21","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"50","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"Proceedings 17th IEEE Annual Conference on Computational Complexity","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.1109/CCC.2002.1004354","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
引用次数: 50

Abstract

An error-correcting code is said to be locally decodable if a randomized algorithm can recover any single bit of a message by reading only a small number of symbols of a possibly corrupted encoding of the message. Katz and Trevisan (2000) showed that any such code C: {0, 1} /spl rarr/ /spl Sigma//sup m/ with a decoding algorithm that makes at most q probes must satisfy m = /spl Omega/((n/log |/spl Sigma/|)/sup q/(q-1)/). They assumed that the decoding algorithm is non-adaptive, and left open the question of proving similar bounds for adaptive decoders. We improve the results of Katz and Trevisan (2000) in two ways. First, we give a more direct proof of their result. Second, and this is our main result, we prove that m = /spl Omega/((n/log|/spl Sigma/|)/sup q/(q-1)/) even if the decoding algorithm is adaptive. An important ingredient of our proof is a randomized method for smoothing an adaptive decoding algorithm. The main technical tool we employ is the Second Moment Method.
为局部可解码代码提供更好的下界
如果随机算法仅通过读取可能损坏的消息编码的少量符号来恢复消息的任何单个比特,则称纠错代码是局部可解码的。Katz和Trevisan(2000)证明了任何这样的代码C: {0, 1} /spl rarr/ /spl Sigma//sup m/,其译码算法必须满足m = /spl Omega/((n/log |/spl Sigma/|)/sup q/(q-1)/)。他们假设解码算法是非自适应的,并留下了证明自适应解码器的类似边界的问题。我们从两个方面改进了Katz和Trevisan(2000)的结果。首先,我们对他们的结果给出了更直接的证明。其次,这是我们的主要结果,我们证明m = /spl Omega/((n/log|/spl Sigma/|)/sup q/(q-1)/),即使解码算法是自适应的。我们证明的一个重要组成部分是随机化平滑自适应解码算法的方法。我们使用的主要技术工具是二阶矩法。
本文章由计算机程序翻译,如有差异,请以英文原文为准。
求助全文
约1分钟内获得全文 求助全文
来源期刊
自引率
0.00%
发文量
0
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
确定
请完成安全验证×
copy
已复制链接
快去分享给好友吧!
我知道了
右上角分享
点击右上角分享
0
联系我们:info@booksci.cn Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。 Copyright © 2023 布克学术 All rights reserved.
京ICP备2023020795号-1
ghs 京公网安备 11010802042870号
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术官方微信