{"title":"收缩肩部:Gibbs抽样收缩模型后验与保证收敛率","authors":"Akihiko Nishimura, Marc A Suchard","doi":"10.1214/22-ba1308","DOIUrl":null,"url":null,"abstract":"<p><p>Use of continuous shrinkage priors - with a \"spike\" near zero and heavy-tails towards infinity - is an increasingly popular approach to induce sparsity in parameter estimates. When the parameters are only weakly identified by the likelihood, however, the posterior may end up with tails as heavy as the prior, jeopardizing robustness of inference. A natural solution is to \"shrink the shoulders\" of a shrinkage prior by lightening up its tails beyond a reasonable parameter range, yielding a <i>regularized</i> version of the prior. We develop a regularization approach which, unlike previous proposals, preserves computationally attractive structures of original shrinkage priors. We study theoretical properties of the Gibbs sampler on resulting posterior distributions, with emphasis on convergence rates of the Pólya-Gamma Gibbs sampler for sparse logistic regression. Our analysis shows that the proposed regularization leads to geometric ergodicity under a broad range of global-local shrinkage priors. Essentially, the only requirement is for the prior <math><msub><mrow><mi>π</mi></mrow><mrow><mtext>local</mtext></mrow></msub><mo>(</mo><mo>⋅</mo><mo>)</mo></math> on the local scale <math><mi>λ</mi></math> to satisfy <math><msub><mrow><mi>π</mi></mrow><mrow><mtext>local</mtext></mrow></msub><mo>(</mo><mn>0</mn><mo>)</mo><mo><</mo><mo>∞</mo></math>. If <math><msub><mrow><mi>π</mi></mrow><mrow><mtext>local</mtext></mrow></msub><mo>(</mo><mo>⋅</mo><mo>)</mo></math> further satisfies <math><msub><mrow><mtext>lim</mtext></mrow><mrow><mi>λ</mi><mo>→</mo><mn>0</mn></mrow></msub><msub><mrow><mi>π</mi></mrow><mrow><mtext>local</mtext></mrow></msub><mo>(</mo><mi>λ</mi><mo>)</mo><mo>/</mo><msup><mrow><mi>λ</mi></mrow><mrow><mi>a</mi></mrow></msup><mo><</mo><mo>∞</mo></math> for <math><mi>a</mi><mo>></mo><mn>0</mn></math>, as in the case of Bayesian bridge priors, we show the sampler to be uniformly ergodic.</p>","PeriodicalId":55398,"journal":{"name":"Bayesian Analysis","volume":null,"pages":null},"PeriodicalIF":4.9000,"publicationDate":"2023-06-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC11105165/pdf/","citationCount":"0","resultStr":"{\"title\":\"Shrinkage with shrunken shoulders: Gibbs sampling shrinkage model posteriors with guaranteed convergence rates.\",\"authors\":\"Akihiko Nishimura, Marc A Suchard\",\"doi\":\"10.1214/22-ba1308\",\"DOIUrl\":null,\"url\":null,\"abstract\":\"<p><p>Use of continuous shrinkage priors - with a \\\"spike\\\" near zero and heavy-tails towards infinity - is an increasingly popular approach to induce sparsity in parameter estimates. When the parameters are only weakly identified by the likelihood, however, the posterior may end up with tails as heavy as the prior, jeopardizing robustness of inference. A natural solution is to \\\"shrink the shoulders\\\" of a shrinkage prior by lightening up its tails beyond a reasonable parameter range, yielding a <i>regularized</i> version of the prior. We develop a regularization approach which, unlike previous proposals, preserves computationally attractive structures of original shrinkage priors. We study theoretical properties of the Gibbs sampler on resulting posterior distributions, with emphasis on convergence rates of the Pólya-Gamma Gibbs sampler for sparse logistic regression. Our analysis shows that the proposed regularization leads to geometric ergodicity under a broad range of global-local shrinkage priors. Essentially, the only requirement is for the prior <math><msub><mrow><mi>π</mi></mrow><mrow><mtext>local</mtext></mrow></msub><mo>(</mo><mo>⋅</mo><mo>)</mo></math> on the local scale <math><mi>λ</mi></math> to satisfy <math><msub><mrow><mi>π</mi></mrow><mrow><mtext>local</mtext></mrow></msub><mo>(</mo><mn>0</mn><mo>)</mo><mo><</mo><mo>∞</mo></math>. If <math><msub><mrow><mi>π</mi></mrow><mrow><mtext>local</mtext></mrow></msub><mo>(</mo><mo>⋅</mo><mo>)</mo></math> further satisfies <math><msub><mrow><mtext>lim</mtext></mrow><mrow><mi>λ</mi><mo>→</mo><mn>0</mn></mrow></msub><msub><mrow><mi>π</mi></mrow><mrow><mtext>local</mtext></mrow></msub><mo>(</mo><mi>λ</mi><mo>)</mo><mo>/</mo><msup><mrow><mi>λ</mi></mrow><mrow><mi>a</mi></mrow></msup><mo><</mo><mo>∞</mo></math> for <math><mi>a</mi><mo>></mo><mn>0</mn></math>, as in the case of Bayesian bridge priors, we show the sampler to be uniformly ergodic.</p>\",\"PeriodicalId\":55398,\"journal\":{\"name\":\"Bayesian Analysis\",\"volume\":null,\"pages\":null},\"PeriodicalIF\":4.9000,\"publicationDate\":\"2023-06-01\",\"publicationTypes\":\"Journal Article\",\"fieldsOfStudy\":null,\"isOpenAccess\":false,\"openAccessPdf\":\"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC11105165/pdf/\",\"citationCount\":\"0\",\"resultStr\":null,\"platform\":\"Semanticscholar\",\"paperid\":null,\"PeriodicalName\":\"Bayesian Analysis\",\"FirstCategoryId\":\"100\",\"ListUrlMain\":\"https://doi.org/10.1214/22-ba1308\",\"RegionNum\":2,\"RegionCategory\":\"数学\",\"ArticlePicture\":[],\"TitleCN\":null,\"AbstractTextCN\":null,\"PMCID\":null,\"EPubDate\":\"2022/4/5 0:00:00\",\"PubModel\":\"Epub\",\"JCR\":\"Q1\",\"JCRName\":\"MATHEMATICS, INTERDISCIPLINARY APPLICATIONS\",\"Score\":null,\"Total\":0}","platform":"Semanticscholar","paperid":null,"PeriodicalName":"Bayesian Analysis","FirstCategoryId":"100","ListUrlMain":"https://doi.org/10.1214/22-ba1308","RegionNum":2,"RegionCategory":"数学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"2022/4/5 0:00:00","PubModel":"Epub","JCR":"Q1","JCRName":"MATHEMATICS, INTERDISCIPLINARY APPLICATIONS","Score":null,"Total":0}
Shrinkage with shrunken shoulders: Gibbs sampling shrinkage model posteriors with guaranteed convergence rates.
Use of continuous shrinkage priors - with a "spike" near zero and heavy-tails towards infinity - is an increasingly popular approach to induce sparsity in parameter estimates. When the parameters are only weakly identified by the likelihood, however, the posterior may end up with tails as heavy as the prior, jeopardizing robustness of inference. A natural solution is to "shrink the shoulders" of a shrinkage prior by lightening up its tails beyond a reasonable parameter range, yielding a regularized version of the prior. We develop a regularization approach which, unlike previous proposals, preserves computationally attractive structures of original shrinkage priors. We study theoretical properties of the Gibbs sampler on resulting posterior distributions, with emphasis on convergence rates of the Pólya-Gamma Gibbs sampler for sparse logistic regression. Our analysis shows that the proposed regularization leads to geometric ergodicity under a broad range of global-local shrinkage priors. Essentially, the only requirement is for the prior on the local scale to satisfy . If further satisfies for , as in the case of Bayesian bridge priors, we show the sampler to be uniformly ergodic.
期刊介绍:
Bayesian Analysis is an electronic journal of the International Society for Bayesian Analysis. It seeks to publish a wide range of articles that demonstrate or discuss Bayesian methods in some theoretical or applied context. The journal welcomes submissions involving presentation of new computational and statistical methods; critical reviews and discussions of existing approaches; historical perspectives; description of important scientific or policy application areas; case studies; and methods for experimental design, data collection, data sharing, or data mining.
Evaluation of submissions is based on importance of content and effectiveness of communication. Discussion papers are typically chosen by the Editor in Chief, or suggested by an Editor, among the regular submissions. In addition, the Journal encourages individual authors to submit manuscripts for consideration as discussion papers.