缺失的 g-质量调查分布的缺失部分

IF 2.2 3区 计算机科学 Q3 COMPUTER SCIENCE, INFORMATION SYSTEMS
Prafulla Chandra;Andrew Thangaraj
{"title":"缺失的 g-质量调查分布的缺失部分","authors":"Prafulla Chandra;Andrew Thangaraj","doi":"10.1109/TIT.2024.3440661","DOIUrl":null,"url":null,"abstract":"Estimating the underlying distribution from iid samples is a classical and important problem in statistics. When the alphabet size is large compared to number of samples, a portion of the distribution is highly likely to be unobserved or sparsely observed. The missing mass, defined as the sum of probabilities \n<inline-formula> <tex-math>$\\Pr (x)$ </tex-math></inline-formula>\n over the missing letters x, and the Good-Turing estimator for missing mass have been important tools in large-alphabet distribution estimation. In this article, given a positive function g from \n<inline-formula> <tex-math>$[{0,1}]$ </tex-math></inline-formula>\n to the reals, the missing g-mass, defined as the sum of \n<inline-formula> <tex-math>$g(\\Pr (x))$ </tex-math></inline-formula>\n over the missing letters x, is introduced and studied. The missing g-mass can be used to investigate the structure of the missing part of the distribution. Specific applications for special cases such as order-\n<inline-formula> <tex-math>$\\alpha $ </tex-math></inline-formula>\n missing mass (\n<inline-formula> <tex-math>$g(p)=p^{\\alpha }$ </tex-math></inline-formula>\n) and the missing Shannon entropy (\n<inline-formula> <tex-math>$g(p)=-p\\log p$ </tex-math></inline-formula>\n) include estimating distance from uniformity of the missing distribution and its partial estimation. Minimax estimation is studied for order-\n<inline-formula> <tex-math>$\\alpha $ </tex-math></inline-formula>\n missing mass for integer values of \n<inline-formula> <tex-math>$\\alpha $ </tex-math></inline-formula>\n and exact minimax convergence rates are obtained. Concentration is studied for a class of functions g and specific results are derived for order-\n<inline-formula> <tex-math>$\\alpha $ </tex-math></inline-formula>\n missing mass and missing Shannon entropy. Sub-Gaussian tail bounds with near-optimal worst-case variance factors are derived. Two new notions of concentration, named strongly sub-Gamma and filtered sub-Gaussian concentration, are introduced and shown to result in right tail bounds that are better than those obtained from sub-Gaussian concentration.","PeriodicalId":13494,"journal":{"name":"IEEE Transactions on Information Theory","volume":"70 10","pages":"7049-7065"},"PeriodicalIF":2.2000,"publicationDate":"2024-08-08","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":"{\"title\":\"Missing g-Mass: Investigating the Missing Parts of Distributions\",\"authors\":\"Prafulla Chandra;Andrew Thangaraj\",\"doi\":\"10.1109/TIT.2024.3440661\",\"DOIUrl\":null,\"url\":null,\"abstract\":\"Estimating the underlying distribution from iid samples is a classical and important problem in statistics. When the alphabet size is large compared to number of samples, a portion of the distribution is highly likely to be unobserved or sparsely observed. The missing mass, defined as the sum of probabilities \\n<inline-formula> <tex-math>$\\\\Pr (x)$ </tex-math></inline-formula>\\n over the missing letters x, and the Good-Turing estimator for missing mass have been important tools in large-alphabet distribution estimation. In this article, given a positive function g from \\n<inline-formula> <tex-math>$[{0,1}]$ </tex-math></inline-formula>\\n to the reals, the missing g-mass, defined as the sum of \\n<inline-formula> <tex-math>$g(\\\\Pr (x))$ </tex-math></inline-formula>\\n over the missing letters x, is introduced and studied. The missing g-mass can be used to investigate the structure of the missing part of the distribution. Specific applications for special cases such as order-\\n<inline-formula> <tex-math>$\\\\alpha $ </tex-math></inline-formula>\\n missing mass (\\n<inline-formula> <tex-math>$g(p)=p^{\\\\alpha }$ </tex-math></inline-formula>\\n) and the missing Shannon entropy (\\n<inline-formula> <tex-math>$g(p)=-p\\\\log p$ </tex-math></inline-formula>\\n) include estimating distance from uniformity of the missing distribution and its partial estimation. Minimax estimation is studied for order-\\n<inline-formula> <tex-math>$\\\\alpha $ </tex-math></inline-formula>\\n missing mass for integer values of \\n<inline-formula> <tex-math>$\\\\alpha $ </tex-math></inline-formula>\\n and exact minimax convergence rates are obtained. Concentration is studied for a class of functions g and specific results are derived for order-\\n<inline-formula> <tex-math>$\\\\alpha $ </tex-math></inline-formula>\\n missing mass and missing Shannon entropy. Sub-Gaussian tail bounds with near-optimal worst-case variance factors are derived. Two new notions of concentration, named strongly sub-Gamma and filtered sub-Gaussian concentration, are introduced and shown to result in right tail bounds that are better than those obtained from sub-Gaussian concentration.\",\"PeriodicalId\":13494,\"journal\":{\"name\":\"IEEE Transactions on Information Theory\",\"volume\":\"70 10\",\"pages\":\"7049-7065\"},\"PeriodicalIF\":2.2000,\"publicationDate\":\"2024-08-08\",\"publicationTypes\":\"Journal Article\",\"fieldsOfStudy\":null,\"isOpenAccess\":false,\"openAccessPdf\":\"\",\"citationCount\":\"0\",\"resultStr\":null,\"platform\":\"Semanticscholar\",\"paperid\":null,\"PeriodicalName\":\"IEEE Transactions on Information Theory\",\"FirstCategoryId\":\"94\",\"ListUrlMain\":\"https://ieeexplore.ieee.org/document/10630859/\",\"RegionNum\":3,\"RegionCategory\":\"计算机科学\",\"ArticlePicture\":[],\"TitleCN\":null,\"AbstractTextCN\":null,\"PMCID\":null,\"EPubDate\":\"\",\"PubModel\":\"\",\"JCR\":\"Q3\",\"JCRName\":\"COMPUTER SCIENCE, INFORMATION SYSTEMS\",\"Score\":null,\"Total\":0}","platform":"Semanticscholar","paperid":null,"PeriodicalName":"IEEE Transactions on Information Theory","FirstCategoryId":"94","ListUrlMain":"https://ieeexplore.ieee.org/document/10630859/","RegionNum":3,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"Q3","JCRName":"COMPUTER SCIENCE, INFORMATION SYSTEMS","Score":null,"Total":0}
引用次数: 0

摘要

从 iid 样本中估计基本分布是统计学中一个经典而重要的问题。当字母表的大小与样本数量相比较大时,分布的一部分极有可能未被观测到或观测稀少。缺失质量定义为缺失字母 x 的概率总和 $\Pr (x)$,缺失质量的 Good-Turing 估计器一直是大字母分布估计的重要工具。本文介绍并研究了从$[{0,1}]$到实数的正函数g的缺失g-质量,其定义为缺失字母x上的$g(\Pr (x))$之和。缺失 g 质量可用于研究分布中缺失部分的结构。特殊情况下的具体应用,如阶- $\alpha $缺失质量($g(p)=p^{\alpha }$)和缺失香农熵($g(p)=-p\log p$),包括估计缺失分布的均匀性距离及其部分估计。针对 $\alpha $ 的整数值,研究了阶 $\alpha $ 缺失质量的最小估计,并获得了精确的最小收敛率。研究了一类函数 g 的集中性,并得出了阶(order- $\alpha $)缺失质量和缺失香农熵的具体结果。推导出了接近最优最坏情况方差系数的亚高斯尾边界。引入了两个新的集中概念,分别称为强亚伽马集中和滤波亚高斯集中,并证明这两个概念能得到比亚高斯集中更好的右尾边界。
本文章由计算机程序翻译,如有差异,请以英文原文为准。
Missing g-Mass: Investigating the Missing Parts of Distributions
Estimating the underlying distribution from iid samples is a classical and important problem in statistics. When the alphabet size is large compared to number of samples, a portion of the distribution is highly likely to be unobserved or sparsely observed. The missing mass, defined as the sum of probabilities $\Pr (x)$ over the missing letters x, and the Good-Turing estimator for missing mass have been important tools in large-alphabet distribution estimation. In this article, given a positive function g from $[{0,1}]$ to the reals, the missing g-mass, defined as the sum of $g(\Pr (x))$ over the missing letters x, is introduced and studied. The missing g-mass can be used to investigate the structure of the missing part of the distribution. Specific applications for special cases such as order- $\alpha $ missing mass ( $g(p)=p^{\alpha }$ ) and the missing Shannon entropy ( $g(p)=-p\log p$ ) include estimating distance from uniformity of the missing distribution and its partial estimation. Minimax estimation is studied for order- $\alpha $ missing mass for integer values of $\alpha $ and exact minimax convergence rates are obtained. Concentration is studied for a class of functions g and specific results are derived for order- $\alpha $ missing mass and missing Shannon entropy. Sub-Gaussian tail bounds with near-optimal worst-case variance factors are derived. Two new notions of concentration, named strongly sub-Gamma and filtered sub-Gaussian concentration, are introduced and shown to result in right tail bounds that are better than those obtained from sub-Gaussian concentration.
求助全文
通过发布文献求助,成功后即可免费获取论文全文。 去求助
来源期刊
IEEE Transactions on Information Theory
IEEE Transactions on Information Theory 工程技术-工程:电子与电气
CiteScore
5.70
自引率
20.00%
发文量
514
审稿时长
12 months
期刊介绍: The IEEE Transactions on Information Theory is a journal that publishes theoretical and experimental papers concerned with the transmission, processing, and utilization of information. The boundaries of acceptable subject matter are intentionally not sharply delimited. Rather, it is hoped that as the focus of research activity changes, a flexible policy will permit this Transactions to follow suit. Current appropriate topics are best reflected by recent Tables of Contents; they are summarized in the titles of editorial areas that appear on the inside front cover.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
确定
请完成安全验证×
copy
已复制链接
快去分享给好友吧!
我知道了
右上角分享
点击右上角分享
0
联系我们:info@booksci.cn Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。 Copyright © 2023 布克学术 All rights reserved.
京ICP备2023020795号-1
ghs 京公网安备 11010802042870号
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术官方微信