{"title":"Probability Mass Functions","authors":"T. Donovan, R. Mickey","doi":"10.1093/OSO/9780198841296.003.0008","DOIUrl":"https://doi.org/10.1093/OSO/9780198841296.003.0008","url":null,"abstract":"This chapter focuses on probability mass functions. One of the primary uses of Bayesian inference is to estimate parameters. To do so, it is necessary to first build a good understanding of probability distributions. This chapter introduces the idea of a random variable and presents general concepts associated with probability distributions for discrete random variables. It starts off by discussing the concept of a function and goes on to describe how a random variable is a type of function. The binomial distribution and the Bernoulli distribution are then used as examples of the probability mass functions (pmf’s). The pmfs can be used to specify prior distributions, likelihoods, likelihood profiles and/or posterior distributions in Bayesian inference.","PeriodicalId":285230,"journal":{"name":"Bayesian Statistics for Beginners","volume":"59 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2019-05-23","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"121698790","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
{"title":"The Maple Syrup Problem: The Normal-Normal Conjugate","authors":"T. Donovan, R. Mickey","doi":"10.1093/OSO/9780198841296.003.0012","DOIUrl":"https://doi.org/10.1093/OSO/9780198841296.003.0012","url":null,"abstract":"In this chapter, Bayesian methods are used to estimate the two parameters that identify a normal distribution, μ and σ. Many Bayesian analyses consider alternative parameter values as hypotheses. The prior distribution for an unknown parameter can be represented by a continuous probability density function when the number of hypotheses is infinite. In the “Maple Syrup Problem,” a normal distribution is used as the prior distribution of μ, the mean number of millions of gallons of maple syrup produced in Vermont in a year. The amount of syrup produced in multiple years is determined, and assumed to follow a normal distribution with known σ. The prior distribution is updated to the posterior distribution in light of this new information. In short, a normal prior distribution + normally distributed data → normal posterior distribution.","PeriodicalId":285230,"journal":{"name":"Bayesian Statistics for Beginners","volume":"12 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2019-05-23","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"115313607","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
{"title":"Bayesian Inference","authors":"T. Donovan, R. Mickey","doi":"10.1093/oso/9780198841296.003.0004","DOIUrl":"https://doi.org/10.1093/oso/9780198841296.003.0004","url":null,"abstract":"Chapter 4 introduces the concept of Bayesian inference. The chapter discusses the scientific method, and illustrates how Bayes’ Theorem can be used for scientific inference. Bayesian inference is the use of Bayes’ Theorem to draw conclusions about a set of mutually exclusive and exhaustive alternative hypotheses by linking prior knowledge about each hypothesis with new data. The result is updated probabilities for each hypothesis of interest. By the end of this chapter, the reader will understand the concepts of induction and deduction, prior probability of a hypothesis, likelihood of the observed data, and posterior probability of a hypothesis, given the data.","PeriodicalId":285230,"journal":{"name":"Bayesian Statistics for Beginners","volume":"1 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2019-05-23","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"123543480","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
{"title":"The White House Problem: The Beta-Binomial Conjugate","authors":"T. Donovan, R. Mickey","doi":"10.1093/OSO/9780198841296.003.0010","DOIUrl":"https://doi.org/10.1093/OSO/9780198841296.003.0010","url":null,"abstract":"This chapter introduces the beta-binomial conjugate. There are special cases where a Bayesian prior probability distribution for an unknown parameter of interest can be quickly updated to a posterior distribution of the same form as the prior. In the “White House Problem,” a beta distribution is used to set the priors for all hypotheses of p, the probability that a famous person can get into the White House without an invitation. Binomial data are then collected, and provide the number of times a famous person gained entry out of a fixed number of attempts. The prior distribution is updated to a posterior distribution (also a beta distribution) in light of this new information. In short, a beta prior distribution for the unknown parameter + binomial data → beta posterior distribution for the unknown parameter, p. The beta distribution is said to be “conjugate to” the binomial distribution.","PeriodicalId":285230,"journal":{"name":"Bayesian Statistics for Beginners","volume":"2 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2019-05-23","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"122263778","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
{"title":"The Survivor Problem: Simple Linear Regression with MCMC","authors":"T. Donovan, R. Mickey","doi":"10.1093/OSO/9780198841296.003.0017","DOIUrl":"https://doi.org/10.1093/OSO/9780198841296.003.0017","url":null,"abstract":"While one of the most common uses of Bayes’ Theorem is in the statistical analysis of a dataset (i.e., statistical modeling), this chapter examines another application of Gibbs sampling: parameter estimation for simple linear regression. In the “Survivor Problem,” the chapter considers the relationship between how many days a contestant lasts in a reality-show competition as a function of how many years of formal education they have. This chapter is a bit more complicated than the previous chapter because it involves estimation of the joint posterior distribution of three parameters. As in earlier chapters, the estimation process is described in detail on a step-by-step basis. Finally, the posterior predictive distribution is estimated and discussed. By the end of the chapter, the reader will have a firm understanding of the following concepts: linear equation, sums of squares, posterior predictive distribution, and linear regression with Markov Chain Monte Carlo and Gibbs sampling.","PeriodicalId":285230,"journal":{"name":"Bayesian Statistics for Beginners","volume":"52 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2019-05-23","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"130980032","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
{"title":"The Portrait Problem: Bayesian Inference with Joint Likelihood","authors":"T. Donovan, R. Mickey","doi":"10.1093/OSO/9780198841296.003.0007","DOIUrl":"https://doi.org/10.1093/OSO/9780198841296.003.0007","url":null,"abstract":"Chapter 7 discusses the “Portrait Problem,” which concerns the dispute about whether a portrait frequently associated with Thomas Bayes (and used, in fact, as the cover of this book!) is actually a picture of him. In doing so, the chapter highlights the fact that multiple pieces of information can be used in a Bayesian analysis. A key concept in this chapter is that multiple sources of data can be combined in a Bayesian inference framework. The main take-home point is that Bayesian analysis can be very, very flexible. A Bayesian analysis is possible as long as the likelihood of observing the data under each hypothesis can be computed. The chapter also discusses the concepts of joint likelihood and independence.","PeriodicalId":285230,"journal":{"name":"Bayesian Statistics for Beginners","volume":"33 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2019-05-23","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"133415119","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
{"title":"MCMC Diagnostic Approaches","authors":"T. Donovan, R. Mickey","doi":"10.1093/OSO/9780198841296.003.0014","DOIUrl":"https://doi.org/10.1093/OSO/9780198841296.003.0014","url":null,"abstract":"The purpose of this chapter is to illustrate some of the things that can go wrong in Markov Chain Monte Carlo (MCMC) analysis and to introduce some diagnostic tools that help identify whether the results of such an analysis can be trusted. The goal of a Bayesian MCMC analysis is to estimate the posterior distribution while skipping the integration required in the denominator of Bayes’ Theorem. The MCMC approach does this by breaking the problem into small, bite-sized pieces, allowing the posterior distribution to be built bit by bit. The main challenge, however, is that several things might go wrong in the process. Several diagnostic tests can be applied to ensure that an MCMC analysis provides an adequate estimate of the posterior distribution. Such diagnostics are required of all MCMC analyses and include tuning, burn-in, and pruning.","PeriodicalId":285230,"journal":{"name":"Bayesian Statistics for Beginners","volume":"106 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2019-05-23","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"127110384","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
{"title":"The Shark Attack Problem Revisited: MCMC with the Metropolis Algorithm","authors":"T. Donovan, R. Mickey","doi":"10.1093/OSO/9780198841296.003.0013","DOIUrl":"https://doi.org/10.1093/OSO/9780198841296.003.0013","url":null,"abstract":"In this chapter, the “Shark Attack Problem” (Chapter 11) is revisited. Markov Chain Monte Carlo (MCMC) is introduced as another way to determine a posterior distribution of λ, the mean number of shark attacks per year. The MCMC approach is so versatile that it can be used to solve almost any kind of parameter estimation problem. The chapter highlights the Metropolis algorithm in detail and illustrates its application, step by step, for the “Shark Attack Problem.” The posterior distribution generated in Chapter 11 using the gamma-Poisson conjugate is compared with the MCMC posterior distribution to show how successful the MCMC method can be. By the end of the chapter, the reader should also understand the following concepts: tuning parameter, MCMC inference, traceplot, and moment matching.","PeriodicalId":285230,"journal":{"name":"Bayesian Statistics for Beginners","volume":"36 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2019-05-23","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"134402883","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
{"title":"The Birthday Problem: Bayesian Inference with Multiple Discrete Hypotheses","authors":"T. Donovan, R. Mickey","doi":"10.1093/OSO/9780198841296.003.0006","DOIUrl":"https://doi.org/10.1093/OSO/9780198841296.003.0006","url":null,"abstract":"The “Birthday Problem” expands consideration from two hypotheses to multiple, discrete hypotheses. In this chapter, interest is in determining the posterior probability that a woman named Mary was born in a given month; there are twelve alternative hypotheses. Furthermore, consideration is given to assigning prior probabilities. The priors represent a priori probabilities that each alternative hypothesis is correct, where a priori means “prior to data collection,” and can be “informative” or “non-informative.” A Bayesian analysis cannot be conducted without using a prior distribution, whether that is an informative prior distribution or a non-informative prior distribution. The chapter discusses objective priors, subjective priors, and prior sensitivity analysis. In addition, the concept of likelihood is explored more deeply.","PeriodicalId":285230,"journal":{"name":"Bayesian Statistics for Beginners","volume":"2012 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2019-05-23","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"131937506","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}