Bayesian Estimation: Information-Theoretic Analysis and Applications

B. Clarke
{"title":"Bayesian Estimation: Information-Theoretic Analysis and Applications","authors":"B. Clarke","doi":"10.1109/ITW.1989.761422","DOIUrl":null,"url":null,"abstract":"It is known that the optimal redundancy of a source code behaves like one half the dimension of the parameter times the log of the sample size. We give an asymptotic expression for the redundancy which is valid for smooth parametric families of distributions equipped with a prior. Important terms include one half the logarithm of the determinant of the Fisher information matrix, minus the logarithm of the prior density and a constant arising from the mean of a Chi-square distribution. The dominant terms arise from an integration by Laplace's method. Our formula can be integrated with respect to the prior distribution on the parameter, under some conditions, so as to give the average redundancy and the minimax redundancy. The minimax code uses Jeff reys' prior. The same expansion has implications for channel coding: Consider channels which have a continuous d-dimensional input alphabet and a k-dimensional output alphabet (where the coordinates of the output are conditionally independent of the input). A message for this channel is Cooperatively encoded by d transmitters and cooperatively decoded by n receivers. For a large number of receivers the mutual information behaves like (d/ 2) logk , that is, one half the number of transmitters times the log of the number of receivers. From the estimation standpoint we have approximated three forms of the cumulative risk under relative entropy loss. Our asymptotic expansions give the risk, the Bayes risk, and the minimax risk. The cumulative risk of the Bayes estimator occurs naturally as the error exponent in a hypothesis test. It also occurs naturally in proving that the standardized posterior converges to a normal.","PeriodicalId":413028,"journal":{"name":"IEEE/CAM Information Theory Workshop at Cornell","volume":"151 1","pages":"0"},"PeriodicalIF":0.0000,"publicationDate":"1989-06-25","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"IEEE/CAM Information Theory Workshop at Cornell","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.1109/ITW.1989.761422","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
引用次数: 0

Abstract

It is known that the optimal redundancy of a source code behaves like one half the dimension of the parameter times the log of the sample size. We give an asymptotic expression for the redundancy which is valid for smooth parametric families of distributions equipped with a prior. Important terms include one half the logarithm of the determinant of the Fisher information matrix, minus the logarithm of the prior density and a constant arising from the mean of a Chi-square distribution. The dominant terms arise from an integration by Laplace's method. Our formula can be integrated with respect to the prior distribution on the parameter, under some conditions, so as to give the average redundancy and the minimax redundancy. The minimax code uses Jeff reys' prior. The same expansion has implications for channel coding: Consider channels which have a continuous d-dimensional input alphabet and a k-dimensional output alphabet (where the coordinates of the output are conditionally independent of the input). A message for this channel is Cooperatively encoded by d transmitters and cooperatively decoded by n receivers. For a large number of receivers the mutual information behaves like (d/ 2) logk , that is, one half the number of transmitters times the log of the number of receivers. From the estimation standpoint we have approximated three forms of the cumulative risk under relative entropy loss. Our asymptotic expansions give the risk, the Bayes risk, and the minimax risk. The cumulative risk of the Bayes estimator occurs naturally as the error exponent in a hypothesis test. It also occurs naturally in proving that the standardized posterior converges to a normal.
贝叶斯估计:信息论分析与应用
众所周知,源代码的最优冗余表现为参数维数的一半乘以样本大小的对数。我们给出了冗余的渐近表达式,该表达式适用于具有先验的光滑参数族分布。重要的项包括二分之一的费雪信息矩阵行列式的对数,减去先验密度的对数和由卡方分布的平均值产生的常数。主导项来自拉普拉斯积分法。在一定条件下,我们的公式可以对参数的先验分布进行积分,从而得到平均冗余和极大最小冗余。极大极小代码使用Jeff reys的先验。同样的扩展对通道编码也有影响:考虑具有连续的d维输入字母表和k维输出字母表的通道(其中输出的坐标与输入条件无关)。该信道的消息由d个发送器协作编码,并由n个接收器协作解码。对于大量的接收器,互信息表现为(d/ 2) logk,即发射器数量的一半乘以接收器数量的对数。从估计的角度出发,我们近似地得到了相对熵损失下累积风险的三种形式。我们的渐近展开式给出了风险,贝叶斯风险和极大极小风险。在假设检验中,贝叶斯估计器的累积风险自然地以误差指数的形式出现。在证明标准化后向法线收敛时也会很自然地出现。
本文章由计算机程序翻译,如有差异,请以英文原文为准。
求助全文
约1分钟内获得全文 求助全文
来源期刊
自引率
0.00%
发文量
0
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
确定
请完成安全验证×
copy
已复制链接
快去分享给好友吧!
我知道了
右上角分享
点击右上角分享
0
联系我们:info@booksci.cn Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。 Copyright © 2023 布克学术 All rights reserved.
京ICP备2023020795号-1
ghs 京公网安备 11010802042870号
Book学术文献互助
Book学术文献互助群
群 号:604180095
Book学术官方微信