The Maximum Entropy of a Metric Space

IF 0.6 4区 数学 Q3 MATHEMATICS
Tom Leinster;Emily Roff
{"title":"The Maximum Entropy of a Metric Space","authors":"Tom Leinster;Emily Roff","doi":"10.1093/qmath/haab003","DOIUrl":null,"url":null,"abstract":"We define a one-parameter family of entropies, each assigning a real number to any probability measure on a compact metric space (or, more generally, a compact Hausdorff space with a notion of similarity between points). These entropies generalise the Shannon and Renyi entropies of information theory. We prove that on any space X, there is a single probability measure maximising all these entropies simultaneously. Moreover, all the entropies have the same maximum value: the maximum entropy of X. As X is scaled up, the maximum entropy grows; its asymptotics determine geometric information about X, including the volume and dimension. We also study the large-scale limit of the maximising measure itself, arguing that it should be regarded as the canonical or uniform measure on X. Primarily we work not with entropy itself but its exponential, called diversity and (in its finite form) used as a measure of biodiversity. Our main theorem was first proved in the finite case by Leinster and Meckes.","PeriodicalId":54522,"journal":{"name":"Quarterly Journal of Mathematics","volume":null,"pages":null},"PeriodicalIF":0.6000,"publicationDate":"2021-01-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://ieeexplore.ieee.org/iel7/8016816/9690900/09690907.pdf","citationCount":"10","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"Quarterly Journal of Mathematics","FirstCategoryId":"100","ListUrlMain":"https://ieeexplore.ieee.org/document/9690907/","RegionNum":4,"RegionCategory":"数学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"Q3","JCRName":"MATHEMATICS","Score":null,"Total":0}
引用次数: 10

Abstract

We define a one-parameter family of entropies, each assigning a real number to any probability measure on a compact metric space (or, more generally, a compact Hausdorff space with a notion of similarity between points). These entropies generalise the Shannon and Renyi entropies of information theory. We prove that on any space X, there is a single probability measure maximising all these entropies simultaneously. Moreover, all the entropies have the same maximum value: the maximum entropy of X. As X is scaled up, the maximum entropy grows; its asymptotics determine geometric information about X, including the volume and dimension. We also study the large-scale limit of the maximising measure itself, arguing that it should be regarded as the canonical or uniform measure on X. Primarily we work not with entropy itself but its exponential, called diversity and (in its finite form) used as a measure of biodiversity. Our main theorem was first proved in the finite case by Leinster and Meckes.
度量空间的最大熵
我们定义了一个单参数熵族,每个熵族为紧致度量空间(或者更一般地说,具有点间相似性概念的紧致Hausdorff空间)上的任何概率测度赋一个实数。这些概括了信息论的香农熵和雷姆尼熵。我们证明了在任意空间X上,存在一个概率度量同时最大化所有这些熵。而且,所有的熵都有相同的最大值:X的最大熵。随着X的增大,最大熵增大,它的渐近性决定了X的几何信息,包括体积和维数。最大化测度本身的大尺度极限提供了一个问题的答案:度量空间上的标准测度是什么?首先,我们研究的不是熵本身,而是它的指数,它的有限形式已经被用来衡量生物多样性。我们的主要定理首先是由伦斯特和梅克斯在有限情况下证明的。
本文章由计算机程序翻译,如有差异,请以英文原文为准。
求助全文
约1分钟内获得全文 求助全文
来源期刊
CiteScore
1.30
自引率
0.00%
发文量
36
审稿时长
6-12 weeks
期刊介绍: The Quarterly Journal of Mathematics publishes original contributions to pure mathematics. All major areas of pure mathematics are represented on the editorial board.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
确定
请完成安全验证×
copy
已复制链接
快去分享给好友吧!
我知道了
右上角分享
点击右上角分享
0
联系我们:info@booksci.cn Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。 Copyright © 2023 布克学术 All rights reserved.
京ICP备2023020795号-1
ghs 京公网安备 11010802042870号
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术官方微信