ST 740: Multiparameter Inference
|
|
- Gwendoline Stewart
- 6 years ago
- Views:
Transcription
1 ST 740: Multiparameter Inference Alyson Wilson Department of Statistics North Carolina State University September 23, 2013 A. Wilson (NCSU Statistics) Multiparameter Inference September 23, / 21
2 Multinomial Model Generalization of the binomial model, for the case where observations can have more than two possible values. Sampling distribution: multinomial with parameters (θ 1,..., θ k ), the probabilities associated to each of the k possible outcomes. Example: In a survey, respondents may: Strongly Agree, Agree, Disagree, Strongly Disagree, or have No Opinion when presented with a statement such as 8am is too early to have class. A. Wilson (NCSU Statistics) Multiparameter Inference September 23, / 21
3 Multinomial Model Sampling Distribution Formally y = (y 1, y 2,..., y k ), a k-vector of counts of the number of observations for each outcome θ j : probability of jth outcome k j=1 θ j = 1 and k j=1 y j = n The sampling distribution has the form f (y θ) Π k j=1θ y j j A. Wilson (NCSU Statistics) Multiparameter Inference September 23, / 21
4 Multinomial Model Conjugate Prior Distribution The conjugate prior distribution for the multinomial sampling distribution is the Dirichlet distribution, which is a multivariate generalization of the beta distribution. with α j > 0 and α 0 = k j=1 α j. π(θ α) k j=1 E[θ j ] = α j α 0 θ α j 1 j Var[θ j ] = α j(α 0 α j ) α0 2(α 0 + 1) Cov(θ i, θ j ) = α iα j α0 2(α 0 + 1) A. Wilson (NCSU Statistics) Multiparameter Inference September 23, / 21
5 Multinomial Model Other Interesting Properties of the Dirichlet Distribution The marginal distribution of any subvector of parameters is also Dirichlet: e.g., (θ i, θ j, 1 θ i θ j ) is Dirichlet(α i, α j, α 0 α i α j ) The marginal distribution of θ j is Beta(α j, α 0 α j ). The marginal distribution of θ i + θ j is Beta(α i + α j, α 0 α i α j ) Common choices for noninformative prior distributions for the Dirichlet are α j = 0, 0.5, or 1, j. α 0 is commonly thought of as the prior sample size. A. Wilson (NCSU Statistics) Multiparameter Inference September 23, / 21
6 Multinomial Model Conjugate Prior Distribution The posterior distribution is also Dirichlet. π(θ α, y) j i=1 θ α j +y j 1 j A. Wilson (NCSU Statistics) Multiparameter Inference September 23, / 21
7 Multinomial Model Example: In a survey, respondents may: Strongly Agree, Agree, Disagree, Strongly Disagree, or have No Opinion when presented with a statement such as 8am is too early to have class. Our sampling distribution is multinomial, with parameters θ SA, θ A, θ D, θ SD, and θ NO, where θ SA + θ A + θ D + θ SD + θ NO = 1. Our prior distribution on the parameters is a noninformative Dirichlet(0.5,0.5,0.5,0.5,0.5), π(θ) θ 0.5 SA θ 0.5 A θ 0.5 SD θ 0.5 D θ 0.5 NO A. Wilson (NCSU Statistics) Multiparameter Inference September 23, / 21
8 Multinomial Model We observe n = 29 responses y = (15, 8, 2, 2, 2). Our posterior distribution is π(θ y) θ SA θ A θ D θ SD θ NO The two probabilities we are interested in are the probability that a student Strongly Agrees that 8am is too early for class, and the probability that a student either Disagrees or Strongly Disagrees that 8am is too early for class. A. Wilson (NCSU Statistics) Multiparameter Inference September 23, / 21
9 Multinomial Model This means that we are interested in the marginal posterior distribution of θ SA, π(θ SA y), and the posterior distribution of θ D + θ SD, π(θ D + θ SD y). We can find these marginal distributions two ways. We can find the marginal distributions analytically using properties of the Dirichlet distribution. In particular, θ SA y Beta(α SA + y SA, α j + n α SA y SA ) = Beta(15.5, 16) θ D + θ SD y Beta(α D + α SD + y D + y SD, αj + n α D α SD y D y SD ) = Beta(5, 26.5) A. Wilson (NCSU Statistics) Multiparameter Inference September 23, / 21
10 Generating Random Samples from Dirichlet Distributions We can simulate directly from the joint posterior Dirichlet distribution (in R, rdirichlet from the LearnBayes package). Gamma method: Two steps for each θ j : 1 Draw x 1, x 2,..., x k from independent Gamma(δ, (α j + y j )) for any common δ 2 Set θ j = x j / k i=1 x i A. Wilson (NCSU Statistics) Multiparameter Inference September 23, / 21
11 Marginal Posterior Distributions Marginal Posterior Posterior P(Strongly Agree) P(Strongly Disagree or Disagree) A. Wilson (NCSU Statistics) Multiparameter Inference September 23, / 21
12 Helpful Hint for Multiparameter Priors A useful idea is that of the induced prior. Suppose that I have a sampling distribution f (y α, β) that depends on two parameters, α and β. However, my prior information is about a function of α and β, say βγ(1 + 1 α ) We want to specify a prior distribution π(α, β) that reflects our prior information. Specifying a prior on π(α, β) induces (specifies) a prior βγ(1 + 1 α ). A. Wilson (NCSU Statistics) Multiparameter Inference September 23, / 21
13 Weibull Parameters alpha <- rgamma(10000,val1,val2) beta <- rgamma(10000,val3,val4) mw <- beta*gamma(1 + 1/alpha) simy <- rweibull(10000,alpha,beta) A. Wilson (NCSU Statistics) Multiparameter Inference September 23, / 21
14 Constraints on the Parameters Start with an unconstrained problem and Bayes Theorem. π(θ y) = f (y θ)π(θ) f (y θ)π(θ)dθ Suppose we want to find the posterior distribution when θ Θ C, where Θ C π(θ)dθ > 0. Define the constrained prior density as π C (θ) = π(θ) Θ C π(θ)dθ, θ Θ C A. Wilson (NCSU Statistics) Multiparameter Inference September 23, / 21
15 Constraints on the Parameters Start with an unconstrained problem and Bayes Theorem. π(θ y) = f (y θ)π(θ) f (y θ)π(θ)dθ Suppose we want to find the posterior distribution when θ Θ C, where Θ C π(θ)dθ > 0. Define the constrained prior density as π C (θ) = π(θ) Θ C π(θ)dθ, θ Θ C A. Wilson (NCSU Statistics) Multiparameter Inference September 23, / 21
16 Constraints on the Parameters Start with an unconstrained problem and Bayes Theorem. π(θ y) = f (y θ)π(θ) f (y θ)π(θ)dθ Suppose we want to find the posterior distribution when θ Θ C, where Θ C π(θ)dθ > 0. Define the constrained prior density as π C (θ) = π(θ) Θ C π(θ)dθ, θ Θ C A. Wilson (NCSU Statistics) Multiparameter Inference September 23, / 21
17 Constraints on the Parameter Using Bayes Theorem π(θ y, θ Θ C ) = = = = f (y θ)π C (θ) Θ C f (y θ)π C (θ)dθ, θ Θ C f (y θ) Θ C f (y θ) π(θ) Θ π(θ)dθ C f (y θ)π(θ) Θ f (y θ)π(θ)dθ π(θ) Θ π(θ)dθ dθ, θ Θ C C Θ C f (y θ)π(θ) Θ f (y θ)π(θ)dθ dθ, θ Θ C π(θ y) Θ C π(θ y)dθ, θ Θ C General idea: Do the unconstrained problem, truncate, renormalize. A. Wilson (NCSU Statistics) Multiparameter Inference September 23, / 21
18 Constraints on the Parameter Using Bayes Theorem π(θ y, θ Θ C ) = = = = f (y θ)π C (θ) Θ C f (y θ)π C (θ)dθ, θ Θ C f (y θ) Θ C f (y θ) π(θ) Θ π(θ)dθ C f (y θ)π(θ) Θ f (y θ)π(θ)dθ π(θ) Θ π(θ)dθ dθ, θ Θ C C Θ C f (y θ)π(θ) Θ f (y θ)π(θ)dθ dθ, θ Θ C π(θ y) Θ C π(θ y)dθ, θ Θ C General idea: Do the unconstrained problem, truncate, renormalize. A. Wilson (NCSU Statistics) Multiparameter Inference September 23, / 21
19 Constraints on the Parameter Example We are trying to estimate the probability that Barry Bonds hits a home run in We observe 73 home runs in 476 at bats. Our prior distribution for p is Beta(1,5) with p (0, 0.15). p Density A. Wilson (NCSU Statistics) Multiparameter Inference September 23, / 21
20 Constraints on the Parameter Example The unconstrained posterior distribution is Beta(1 + 73, ). p Density A. Wilson (NCSU Statistics) Multiparameter Inference September 23, / 21
21 Less Nice Example Suppose that our sampling distribution is Gamma(α, β) and we choose independent marginal prior distributions π(α) Gamma(2, 1) and π(β) Gamma(5, 1). π(α, β y) βnα Γ(α) n exp( β y i )( y i ) α 1 exp( (α + β))αβ 4 βnα+4 Γ(α) n exp( β( y i + 1))α exp( α)( y i ) α 1 We want to make posterior inferences about α, β, and the predictive distribution for the next observation. A. Wilson (NCSU Statistics) Multiparameter Inference September 23, / 21
22 Predictive Distribution The predictive distribution is the distribution of the next observation conditional on the data observed so far. The uncertainty about the parameters is integrated out. f (y n+1 y) = f (y n+1 α, β)π(α, β y)dαdβ where f (y n+1 α, β) = βα Γ(α) exp( βy n+1)(y n+1 ) α 1 A. Wilson (NCSU Statistics) Multiparameter Inference September 23, / 21
23 Contour Plot of Posterior Distribution β α A. Wilson (NCSU Statistics) Multiparameter Inference September 23, / 21
24 Joint, Marginal, and Conditional Distributions Posterior? π(α, β y) βnα+4 Γ(α) n exp( β( y i + 1))α exp( α)( y i ) α 1 Conditionals? π(α β, y) hard. β α, y Gamma(nα + 5, y i + 1) Marginals? π(β y) hard. π(α y) α exp( α) Γ(α) n Γ(nα+5) ( y i +1) nα+5 ( y i ) α 1 A. Wilson (NCSU Statistics) Multiparameter Inference September 23, / 21
ST 740: Model Selection
ST 740: Model Selection Alyson Wilson Department of Statistics North Carolina State University November 25, 2013 A. Wilson (NCSU Statistics) Model Selection November 25, 2013 1 / 29 Formal Bayesian Model
More informationFoundations of Statistical Inference
Foundations of Statistical Inference Julien Berestycki Department of Statistics University of Oxford MT 2016 Julien Berestycki (University of Oxford) SB2a MT 2016 1 / 20 Lecture 6 : Bayesian Inference
More informationIntroduction to Bayesian Methods
Introduction to Bayesian Methods Jessi Cisewski Department of Statistics Yale University Sagan Summer Workshop 2016 Our goal: introduction to Bayesian methods Likelihoods Priors: conjugate priors, non-informative
More informationA Very Brief Summary of Bayesian Inference, and Examples
A Very Brief Summary of Bayesian Inference, and Examples Trinity Term 009 Prof Gesine Reinert Our starting point are data x = x 1, x,, x n, which we view as realisations of random variables X 1, X,, X
More informationMultinomial Data. f(y θ) θ y i. where θ i is the probability that a given trial results in category i, i = 1,..., k. The parameter space is
Multinomial Data The multinomial distribution is a generalization of the binomial for the situation in which each trial results in one and only one of several categories, as opposed to just two, as in
More informationBayesian Inference: Posterior Intervals
Bayesian Inference: Posterior Intervals Simple values like the posterior mean E[θ X] and posterior variance var[θ X] can be useful in learning about θ. Quantiles of π(θ X) (especially the posterior median)
More informationBeta statistics. Keywords. Bayes theorem. Bayes rule
Keywords Beta statistics Tommy Norberg tommy@chalmers.se Mathematical Sciences Chalmers University of Technology Gothenburg, SWEDEN Bayes s formula Prior density Likelihood Posterior density Conjugate
More informationChapter 5. Bayesian Statistics
Chapter 5. Bayesian Statistics Principles of Bayesian Statistics Anything unknown is given a probability distribution, representing degrees of belief [subjective probability]. Degrees of belief [subjective
More informationIntroduction to Bayesian Methods. Introduction to Bayesian Methods p.1/??
to Bayesian Methods Introduction to Bayesian Methods p.1/?? We develop the Bayesian paradigm for parametric inference. To this end, suppose we conduct (or wish to design) a study, in which the parameter
More informationParameter Estimation
Parameter Estimation Chapters 13-15 Stat 477 - Loss Models Chapters 13-15 (Stat 477) Parameter Estimation Brian Hartman - BYU 1 / 23 Methods for parameter estimation Methods for parameter estimation Methods
More informationProbability Theory. Patrick Lam
Probability Theory Patrick Lam Outline Probability Random Variables Simulation Important Distributions Discrete Distributions Continuous Distributions Most Basic Definition of Probability: number of successes
More informationUSEFUL PROPERTIES OF THE MULTIVARIATE NORMAL*
USEFUL PROPERTIES OF THE MULTIVARIATE NORMAL* 3 Conditionals and marginals For Bayesian analysis it is very useful to understand how to write joint, marginal, and conditional distributions for the multivariate
More informationST 740: Linear Models and Multivariate Normal Inference
ST 740: Linear Models and Multivariate Normal Inference Alyson Wilson Department of Statistics North Carolina State University November 4, 2013 A. Wilson (NCSU STAT) Linear Models November 4, 2013 1 /
More informationInference for a Population Proportion
Al Nosedal. University of Toronto. November 11, 2015 Statistical inference is drawing conclusions about an entire population based on data in a sample drawn from that population. From both frequentist
More informationBayesian inference. Fredrik Ronquist and Peter Beerli. October 3, 2007
Bayesian inference Fredrik Ronquist and Peter Beerli October 3, 2007 1 Introduction The last few decades has seen a growing interest in Bayesian inference, an alternative approach to statistical inference.
More informationHPD Intervals / Regions
HPD Intervals / Regions The HPD region will be an interval when the posterior is unimodal. If the posterior is multimodal, the HPD region might be a discontiguous set. Picture: The set {θ : θ (1.5, 3.9)
More informationBAYESIAN METHODS FOR VARIABLE SELECTION WITH APPLICATIONS TO HIGH-DIMENSIONAL DATA
BAYESIAN METHODS FOR VARIABLE SELECTION WITH APPLICATIONS TO HIGH-DIMENSIONAL DATA Intro: Course Outline and Brief Intro to Marina Vannucci Rice University, USA PASI-CIMAT 04/28-30/2010 Marina Vannucci
More informationLecture 2: Conjugate priors
(Spring ʼ) Lecture : Conjugate priors Julia Hockenmaier juliahmr@illinois.edu Siebel Center http://www.cs.uiuc.edu/class/sp/cs98jhm The binomial distribution If p is the probability of heads, the probability
More informationECE531 Homework Assignment Number 6 Solution
ECE53 Homework Assignment Number 6 Solution Due by 8:5pm on Wednesday 3-Mar- Make sure your reasoning and work are clear to receive full credit for each problem.. 6 points. Suppose you have a scalar random
More informationStat 5101 Lecture Slides: Deck 8 Dirichlet Distribution. Charles J. Geyer School of Statistics University of Minnesota
Stat 5101 Lecture Slides: Deck 8 Dirichlet Distribution Charles J. Geyer School of Statistics University of Minnesota 1 The Dirichlet Distribution The Dirichlet Distribution is to the beta distribution
More informationPrior Choice, Summarizing the Posterior
Prior Choice, Summarizing the Posterior Statistics 220 Spring 2005 Copyright c 2005 by Mark E. Irwin Informative Priors Binomial Model: y π Bin(n, π) π is the success probability. Need prior p(π) Bayes
More informationBayesian Models in Machine Learning
Bayesian Models in Machine Learning Lukáš Burget Escuela de Ciencias Informáticas 2017 Buenos Aires, July 24-29 2017 Frequentist vs. Bayesian Frequentist point of view: Probability is the frequency of
More informationIntroduction to Probabilistic Machine Learning
Introduction to Probabilistic Machine Learning Piyush Rai Dept. of CSE, IIT Kanpur (Mini-course 1) Nov 03, 2015 Piyush Rai (IIT Kanpur) Introduction to Probabilistic Machine Learning 1 Machine Learning
More informationPARAMETER ESTIMATION: BAYESIAN APPROACH. These notes summarize the lectures on Bayesian parameter estimation.
PARAMETER ESTIMATION: BAYESIAN APPROACH. These notes summarize the lectures on Bayesian parameter estimation.. Beta Distribution We ll start by learning about the Beta distribution, since we end up using
More informationSTAT 425: Introduction to Bayesian Analysis
STAT 425: Introduction to Bayesian Analysis Marina Vannucci Rice University, USA Fall 2017 Marina Vannucci (Rice University, USA) Bayesian Analysis (Part 1) Fall 2017 1 / 10 Lecture 7: Prior Types Subjective
More informationSimulating Random Variables
Simulating Random Variables Timothy Hanson Department of Statistics, University of South Carolina Stat 740: Statistical Computing 1 / 23 R has many built-in random number generators... Beta, gamma (also
More informationAlternative prior assumptions for improving the performance of naïve Bayesian classifiers
Data Min Knowl Disc DOI 1.17/s1618-8-11-6 Alternative prior assumptions for improving the performance of naïve Bayesian classifiers Tzu-Tsung Wong Received: 2 July 27 / Accepted: 14 May 28 Springer Science+Business
More informationIntroduction to Applied Bayesian Modeling. ICPSR Day 4
Introduction to Applied Bayesian Modeling ICPSR Day 4 Simple Priors Remember Bayes Law: Where P(A) is the prior probability of A Simple prior Recall the test for disease example where we specified the
More informationStat 5101 Lecture Notes
Stat 5101 Lecture Notes Charles J. Geyer Copyright 1998, 1999, 2000, 2001 by Charles J. Geyer May 7, 2001 ii Stat 5101 (Geyer) Course Notes Contents 1 Random Variables and Change of Variables 1 1.1 Random
More informationCOS513 LECTURE 8 STATISTICAL CONCEPTS
COS513 LECTURE 8 STATISTICAL CONCEPTS NIKOLAI SLAVOV AND ANKUR PARIKH 1. MAKING MEANINGFUL STATEMENTS FROM JOINT PROBABILITY DISTRIBUTIONS. A graphical model (GM) represents a family of probability distributions
More informationConjugate Analysis for the Linear Model
Conjugate Analysis for the Linear Model If we have good prior knowledge that can help us specify priors for β and σ 2, we can use conjugate priors. Following the procedure in Christensen, Johnson, Branscum,
More informationLecture 2: Statistical Decision Theory (Part I)
Lecture 2: Statistical Decision Theory (Part I) Hao Helen Zhang Hao Helen Zhang Lecture 2: Statistical Decision Theory (Part I) 1 / 35 Outline of This Note Part I: Statistics Decision Theory (from Statistical
More informationChapter 5 continued. Chapter 5 sections
Chapter 5 sections Discrete univariate distributions: 5.2 Bernoulli and Binomial distributions Just skim 5.3 Hypergeometric distributions 5.4 Poisson distributions Just skim 5.5 Negative Binomial distributions
More informationEXERCISES FOR SECTION 1 AND 2
EXERCISES FOR SECTION AND Exercise. (Conditional probability). Suppose that if θ, then y has a normal distribution with mean and standard deviation σ, and if θ, then y has a normal distribution with mean
More informationPart III. A Decision-Theoretic Approach and Bayesian testing
Part III A Decision-Theoretic Approach and Bayesian testing 1 Chapter 10 Bayesian Inference as a Decision Problem The decision-theoretic framework starts with the following situation. We would like to
More informationIntroduction to Bayesian Statistics 1
Introduction to Bayesian Statistics 1 STA 442/2101 Fall 2018 1 This slide show is an open-source document. See last slide for copyright information. 1 / 42 Thomas Bayes (1701-1761) Image from the Wikipedia
More informationOne-parameter models
One-parameter models Patrick Breheny January 22 Patrick Breheny BST 701: Bayesian Modeling in Biostatistics 1/17 Introduction Binomial data is not the only example in which Bayesian solutions can be worked
More informationAdvanced topics from statistics
Advanced topics from statistics Anders Ringgaard Kristensen Advanced Herd Management Slide 1 Outline Covariance and correlation Random vectors and multivariate distributions The multinomial distribution
More informationBayes and Empirical Bayes Estimation of the Scale Parameter of the Gamma Distribution under Balanced Loss Functions
The Korean Communications in Statistics Vol. 14 No. 1, 2007, pp. 71 80 Bayes and Empirical Bayes Estimation of the Scale Parameter of the Gamma Distribution under Balanced Loss Functions R. Rezaeian 1)
More informationBayesian inference. Rasmus Waagepetersen Department of Mathematics Aalborg University Denmark. April 10, 2017
Bayesian inference Rasmus Waagepetersen Department of Mathematics Aalborg University Denmark April 10, 2017 1 / 22 Outline for today A genetic example Bayes theorem Examples Priors Posterior summaries
More informationPATTERN RECOGNITION AND MACHINE LEARNING CHAPTER 2: PROBABILITY DISTRIBUTIONS
PATTERN RECOGNITION AND MACHINE LEARNING CHAPTER 2: PROBABILITY DISTRIBUTIONS Parametric Distributions Basic building blocks: Need to determine given Representation: or? Recall Curve Fitting Binary Variables
More informationBayesian Inference and Computation
Southern Illinois University Carbondale OpenSIUC Research Papers Graduate School Fall 11-8-2010 Bayesian Inference and Computation Rasanji C. Rathnayake rasanjirathnayake@yahoo.com Follow this and additional
More informationWeakness of Beta priors (or conjugate priors in general) They can only represent a limited range of prior beliefs. For example... There are no bimodal beta distributions (except when the modes are at 0
More informationBayesian Methods for Data Analysis
. Course contents Bayesian Methods for Data Analysis ENAR Annual Meeting Introduction of Bayesian concepts using single-parameter models. Multiple-parameter models and hyerarchical models. Computation:
More informationBayesian Statistics Part III: Building Bayes Theorem Part IV: Prior Specification
Bayesian Statistics Part III: Building Bayes Theorem Part IV: Prior Specification Michael Anderson, PhD Hélène Carabin, DVM, PhD Department of Biostatistics and Epidemiology The University of Oklahoma
More informationLecture 2: Priors and Conjugacy
Lecture 2: Priors and Conjugacy Melih Kandemir melih.kandemir@iwr.uni-heidelberg.de May 6, 2014 Some nice courses Fred A. Hamprecht (Heidelberg U.) https://www.youtube.com/watch?v=j66rrnzzkow Michael I.
More informationContents. Part I: Fundamentals of Bayesian Inference 1
Contents Preface xiii Part I: Fundamentals of Bayesian Inference 1 1 Probability and inference 3 1.1 The three steps of Bayesian data analysis 3 1.2 General notation for statistical inference 4 1.3 Bayesian
More informationHIERARCHICAL BAYESIAN ANALYSIS OF BINARY MATCHED PAIRS DATA
Statistica Sinica 1(2), 647-657 HIERARCHICAL BAYESIAN ANALYSIS OF BINARY MATCHED PAIRS DATA Malay Ghosh, Ming-Hui Chen, Atalanta Ghosh and Alan Agresti University of Florida, Worcester Polytechnic Institute
More informationA Few Special Distributions and Their Properties
A Few Special Distributions and Their Properties Econ 690 Purdue University Justin L. Tobias (Purdue) Distributional Catalog 1 / 20 Special Distributions and Their Associated Properties 1 Uniform Distribution
More informationA Very Brief Summary of Statistical Inference, and Examples
A Very Brief Summary of Statistical Inference, and Examples Trinity Term 2008 Prof. Gesine Reinert 1 Data x = x 1, x 2,..., x n, realisations of random variables X 1, X 2,..., X n with distribution (model)
More informationBayesian inference: what it means and why we care
Bayesian inference: what it means and why we care Robin J. Ryder Centre de Recherche en Mathématiques de la Décision Université Paris-Dauphine 6 November 2017 Mathematical Coffees Robin Ryder (Dauphine)
More informationWeakness of Beta priors (or conjugate priors in general) They can only represent a limited range of prior beliefs. For example... There are no bimodal beta distributions (except when the modes are at 0
More information9 Bayesian inference. 9.1 Subjective probability
9 Bayesian inference 1702-1761 9.1 Subjective probability This is probability regarded as degree of belief. A subjective probability of an event A is assessed as p if you are prepared to stake pm to win
More informationLearning Bayesian network : Given structure and completely observed data
Learning Bayesian network : Given structure and completely observed data Probabilistic Graphical Models Sharif University of Technology Spring 2017 Soleymani Learning problem Target: true distribution
More informationMultiparameter models (cont.)
Multiparameter models (cont.) Dr. Jarad Niemi STAT 544 - Iowa State University February 1, 2018 Jarad Niemi (STAT544@ISU) Multiparameter models (cont.) February 1, 2018 1 / 20 Outline Multinomial Multivariate
More informationThe binomial model. Assume a uniform prior distribution on p(θ). Write the pdf for this distribution.
The binomial model Example. After suspicious performance in the weekly soccer match, 37 mathematical sciences students, staff, and faculty were tested for the use of performance enhancing analytics. Let
More informationMarkov Chain Monte Carlo methods
Markov Chain Monte Carlo methods Tomas McKelvey and Lennart Svensson Signal Processing Group Department of Signals and Systems Chalmers University of Technology, Sweden November 26, 2012 Today s learning
More informationStatistical Theory MT 2006 Problems 4: Solution sketches
Statistical Theory MT 006 Problems 4: Solution sketches 1. Suppose that X has a Poisson distribution with unknown mean θ. Determine the conjugate prior, and associate posterior distribution, for θ. Determine
More informationBayesian philosophy Bayesian computation Bayesian software. Bayesian Statistics. Petter Mostad. Chalmers. April 6, 2017
Chalmers April 6, 2017 Bayesian philosophy Bayesian philosophy Bayesian statistics versus classical statistics: War or co-existence? Classical statistics: Models have variables and parameters; these are
More informationModule 22: Bayesian Methods Lecture 9 A: Default prior selection
Module 22: Bayesian Methods Lecture 9 A: Default prior selection Peter Hoff Departments of Statistics and Biostatistics University of Washington Outline Jeffreys prior Unit information priors Empirical
More informationStatistics/Mathematics 310 Larget April 14, Solution to Assignment #9
Solution to Assignment #9 1. A sample s = (753.0, 1396.9, 1528.2, 1646.9, 717.6, 446.5, 848.0, 1222.0, 748.5, 610.1) is modeled as an i.i.d sample from a Gamma(α, λ) distribution. You may think of this
More informationLanguage as a Stochastic Process
CS769 Spring 2010 Advanced Natural Language Processing Language as a Stochastic Process Lecturer: Xiaojin Zhu jerryzhu@cs.wisc.edu 1 Basic Statistics for NLP Pick an arbitrary letter x at random from any
More informationPart 3 Robust Bayesian statistics & applications in reliability networks
Tuesday 9:00-12:30 Part 3 Robust Bayesian statistics & applications in reliability networks by Gero Walter 69 Robust Bayesian statistics & applications in reliability networks Outline Robust Bayesian Analysis
More informationDepartment of Large Animal Sciences. Outline. Slide 2. Department of Large Animal Sciences. Slide 4. Department of Large Animal Sciences
Outline Advanced topics from statistics Anders Ringgaard Kristensen Covariance and correlation Random vectors and multivariate distributions The multinomial distribution The multivariate normal distribution
More informationContents 1. Contents
Contents 1 Contents 6 Distributions of Functions of Random Variables 2 6.1 Transformation of Discrete r.v.s............. 3 6.2 Method of Distribution Functions............. 6 6.3 Method of Transformations................
More informationEcon 2140, spring 2018, Part IIa Statistical Decision Theory
Econ 2140, spring 2018, Part IIa Maximilian Kasy Department of Economics, Harvard University 1 / 35 Examples of decision problems Decide whether or not the hypothesis of no racial discrimination in job
More informationeqr094: Hierarchical MCMC for Bayesian System Reliability
eqr094: Hierarchical MCMC for Bayesian System Reliability Alyson G. Wilson Statistical Sciences Group, Los Alamos National Laboratory P.O. Box 1663, MS F600 Los Alamos, NM 87545 USA Phone: 505-667-9167
More informationLecture 6. Prior distributions
Summary Lecture 6. Prior distributions 1. Introduction 2. Bivariate conjugate: normal 3. Non-informative / reference priors Jeffreys priors Location parameters Proportions Counts and rates Scale parameters
More informationChapter 8: Sampling distributions of estimators Sections
Chapter 8: Sampling distributions of estimators Sections 8.1 Sampling distribution of a statistic 8.2 The Chi-square distributions 8.3 Joint Distribution of the sample mean and sample variance Skip: p.
More informationBayesian Statistics Script for the course in spring 2014
Bayesian Statistics Script for the course in spring 2014 Hansruedi Künsch Seminar für Statistik, ETH Zurich Version of August 6, 2014 Contents 1 Basic concepts 1 1.1 Interpretations of probability..........................
More informationBayes: All uncertainty is described using probability.
Bayes: All uncertainty is described using probability. Let w be the data and θ be any unknown quantities. Likelihood. The probability model π(w θ) has θ fixed and w varying. The likelihood L(θ; w) is π(w
More informationIntroduction to Bayesian Statistics with WinBUGS Part 4 Priors and Hierarchical Models
Introduction to Bayesian Statistics with WinBUGS Part 4 Priors and Hierarchical Models Matthew S. Johnson New York ASA Chapter Workshop CUNY Graduate Center New York, NY hspace1in December 17, 2009 December
More informationThe Bayesian Paradigm
Stat 200 The Bayesian Paradigm Friday March 2nd The Bayesian Paradigm can be seen in some ways as an extra step in the modelling world just as parametric modelling is. We have seen how we could use probabilistic
More informationSTAT J535: Chapter 5: Classes of Bayesian Priors
STAT J535: Chapter 5: Classes of Bayesian Priors David B. Hitchcock E-Mail: hitchcock@stat.sc.edu Spring 2012 The Bayesian Prior A prior distribution must be specified in a Bayesian analysis. The choice
More informationChapter 4 HOMEWORK ASSIGNMENTS. 4.1 Homework #1
Chapter 4 HOMEWORK ASSIGNMENTS These homeworks may be modified as the semester progresses. It is your responsibility to keep up to date with the correctly assigned homeworks. There may be some errors in
More information10. Exchangeability and hierarchical models Objective. Recommended reading
10. Exchangeability and hierarchical models Objective Introduce exchangeability and its relation to Bayesian hierarchical models. Show how to fit such models using fully and empirical Bayesian methods.
More informationBayesian statistics, simulation and software
Module 4: Normal model, improper and conjugate priors Department of Mathematical Sciences Aalborg University 1/25 Another example: normal sample with known precision Heights of some Copenhageners in 1995:
More informationBayesian statistics, simulation and software
Module 10: Bayesian prediction and model checking Department of Mathematical Sciences Aalborg University 1/15 Prior predictions Suppose we want to predict future data x without observing any data x. Assume:
More informationMAS3301 Bayesian Statistics
MAS3301 Bayesian Statistics M. Farrow School of Mathematics and Statistics Newcastle University Semester 2, 2008-9 1 11 Conjugate Priors IV: The Dirichlet distribution and multinomial observations 11.1
More informationStats 579 Intermediate Bayesian Modeling. Assignment # 2 Solutions
Stats 579 Intermediate Bayesian Modeling Assignment # 2 Solutions 1. Let w Gy) with y a vector having density fy θ) and G having a differentiable inverse function. Find the density of w in general and
More informationEstimation of reliability parameters from Experimental data (Parte 2) Prof. Enrico Zio
Estimation of reliability parameters from Experimental data (Parte 2) This lecture Life test (t 1,t 2,...,t n ) Estimate θ of f T t θ For example: λ of f T (t)= λe - λt Classical approach (frequentist
More informationStatistical Theory MT 2007 Problems 4: Solution sketches
Statistical Theory MT 007 Problems 4: Solution sketches 1. Consider a 1-parameter exponential family model with density f(x θ) = f(x)g(θ)exp{cφ(θ)h(x)}, x X. Suppose that the prior distribution has the
More informationBayesian Inference. p(y)
Bayesian Inference There are different ways to interpret a probability statement in a real world setting. Frequentist interpretations of probability apply to situations that can be repeated many times,
More informationVCMC: Variational Consensus Monte Carlo
VCMC: Variational Consensus Monte Carlo Maxim Rabinovich, Elaine Angelino, Michael I. Jordan Berkeley Vision and Learning Center September 22, 2015 probabilistic models! sky fog bridge water grass object
More informationINTRODUCTION TO BAYESIAN STATISTICS
INTRODUCTION TO BAYESIAN STATISTICS Sarat C. Dass Department of Statistics & Probability Department of Computer Science & Engineering Michigan State University TOPICS The Bayesian Framework Different Types
More informationProbability and Estimation. Alan Moses
Probability and Estimation Alan Moses Random variables and probability A random variable is like a variable in algebra (e.g., y=e x ), but where at least part of the variability is taken to be stochastic.
More informationBayesian Methods for Estimating the Reliability of Complex Systems Using Heterogeneous Multilevel Information
Statistics Preprints Statistics 8-2010 Bayesian Methods for Estimating the Reliability of Complex Systems Using Heterogeneous Multilevel Information Jiqiang Guo Iowa State University, jqguo@iastate.edu
More informationThe devil is in the denominator
Chapter 6 The devil is in the denominator 6. Too many coin flips Suppose we flip two coins. Each coin i is either fair (P r(h) = θ i = 0.5) or biased towards heads (P r(h) = θ i = 0.9) however, we cannot
More informationMotif representation using position weight matrix
Motif representation using position weight matrix Xiaohui Xie University of California, Irvine Motif representation using position weight matrix p.1/31 Position weight matrix Position weight matrix representation
More informationSTAT Advanced Bayesian Inference
1 / 32 STAT 625 - Advanced Bayesian Inference Meng Li Department of Statistics Jan 23, 218 The Dirichlet distribution 2 / 32 θ Dirichlet(a 1,...,a k ) with density p(θ 1,θ 2,...,θ k ) = k j=1 Γ(a j) Γ(
More informationBayesian Inference for Regression Parameters
Bayesian Inference for Regression Parameters 1 Bayesian inference for simple linear regression parameters follows the usual pattern for all Bayesian analyses: 1. Form a prior distribution over all unknown
More informationIntroduc)on to Bayesian Methods
Introduc)on to Bayesian Methods Bayes Rule py x)px) = px! y) = px y)py) py x) = px y)py) px) px) =! px! y) = px y)py) y py x) = py x) =! y "! y px y)py) px y)py) px y)py) px y)py)dy Bayes Rule py x) =
More informationPMR Learning as Inference
Outline PMR Learning as Inference Probabilistic Modelling and Reasoning Amos Storkey Modelling 2 The Exponential Family 3 Bayesian Sets School of Informatics, University of Edinburgh Amos Storkey PMR Learning
More informationModel comparison. Christopher A. Sims Princeton University October 18, 2016
ECO 513 Fall 2008 Model comparison Christopher A. Sims Princeton University sims@princeton.edu October 18, 2016 c 2016 by Christopher A. Sims. This document may be reproduced for educational and research
More informationData Analysis and Uncertainty Part 2: Estimation
Data Analysis and Uncertainty Part 2: Estimation Instructor: Sargur N. University at Buffalo The State University of New York srihari@cedar.buffalo.edu 1 Topics in Estimation 1. Estimation 2. Desirable
More informationLecture 3. Univariate Bayesian inference: conjugate analysis
Summary Lecture 3. Univariate Bayesian inference: conjugate analysis 1. Posterior predictive distributions 2. Conjugate analysis for proportions 3. Posterior predictions for proportions 4. Conjugate analysis
More informationPOSTERIOR PROPRIETY IN SOME HIERARCHICAL EXPONENTIAL FAMILY MODELS
POSTERIOR PROPRIETY IN SOME HIERARCHICAL EXPONENTIAL FAMILY MODELS EDWARD I. GEORGE and ZUOSHUN ZHANG The University of Texas at Austin and Quintiles Inc. June 2 SUMMARY For Bayesian analysis of hierarchical
More information1 Maximum Likelihood Estimation
heads tails Figure 1: A simple thumbtack tossing experiment. L(θ :D) 0 0.2 0.4 0.6 0.8 1 Figure 2: The likelihood function for the sequence of tosses H,T,T,H,H. 1 Maximum Likelihood Estimation In this
More informationExact distribution theory for belief net responses
Exact distribution theory for belief net responses Peter M. Hooper Department of Mathematical and Statistical Sciences University of Alberta Edmonton, Canada, T6G 2G1 hooper@stat.ualberta.ca May 2, 2008
More informationPattern Recognition and Machine Learning. Bishop Chapter 2: Probability Distributions
Pattern Recognition and Machine Learning Chapter 2: Probability Distributions Cécile Amblard Alex Kläser Jakob Verbeek October 11, 27 Probability Distributions: General Density Estimation: given a finite
More information