Classical and Bayesian inference

Similar documents
Classical and Bayesian inference

Classical and Bayesian inference

Homework 1: Solution

Probability and Statistics

STAT J535: Chapter 5: Classes of Bayesian Priors

Poisson CI s. Robert L. Wolpert Department of Statistical Science Duke University, Durham, NC, USA

Chapter 5. Bayesian Statistics

Beta statistics. Keywords. Bayes theorem. Bayes rule

Estimation of reliability parameters from Experimental data (Parte 2) Prof. Enrico Zio

Introduction to Bayesian Statistics with WinBUGS Part 4 Priors and Hierarchical Models

Bayesian Statistics Part III: Building Bayes Theorem Part IV: Prior Specification

Bayesian Regression Linear and Logistic Regression

Continuous Probability Spaces

Some Asymptotic Bayesian Inference (background to Chapter 2 of Tanner s book)

One Parameter Models

CS 361: Probability & Statistics

One-parameter models

Chapter 8: Sampling distributions of estimators Sections

Remarks on Improper Ignorance Priors

Introduction to Probabilistic Machine Learning

PARAMETER ESTIMATION: BAYESIAN APPROACH. These notes summarize the lectures on Bayesian parameter estimation.

Fundamentals. CS 281A: Statistical Learning Theory. Yangqing Jia. August, Based on tutorial slides by Lester Mackey and Ariel Kleiner

Introduction to Applied Bayesian Modeling. ICPSR Day 4

Parametric Models. Dr. Shuang LIANG. School of Software Engineering TongJi University Fall, 2012

PATTERN RECOGNITION AND MACHINE LEARNING CHAPTER 2: PROBABILITY DISTRIBUTIONS

Bayesian Learning. HT2015: SC4 Statistical Data Mining and Machine Learning. Maximum Likelihood Principle. The Bayesian Learning Framework

Model Checking and Improvement

Computational Perception. Bayesian Inference

Chapter 5. Chapter 5 sections

1 Introduction. P (n = 1 red ball drawn) =

DS-GA 1002 Lecture notes 11 Fall Bayesian statistics

PMR Learning as Inference

Module 22: Bayesian Methods Lecture 9 A: Default prior selection

Statistics 360/601 Modern Bayesian Theory

Statistical Theory MT 2007 Problems 4: Solution sketches

Non-Parametric Bayes

1 A simple example. A short introduction to Bayesian statistics, part I Math 217 Probability and Statistics Prof. D.

Definition A random variable X is said to have a Weibull distribution with parameters α and β (α > 0, β > 0) if the pdf of X is

Theory of Maximum Likelihood Estimation. Konstantin Kashin

CS Lecture 19. Exponential Families & Expectation Propagation

Foundations of Statistical Inference

STAT Advanced Bayesian Inference

Data Analysis and Uncertainty Part 2: Estimation

Math 494: Mathematical Statistics

Probability Distributions Columns (a) through (d)

CS 361: Probability & Statistics

Predictive Distributions

Continuous random variables

Probability and Estimation. Alan Moses

Chapter 5 continued. Chapter 5 sections

(Ch 3.4.1, 3.4.2, 4.1, 4.2, 4.3)

General Bayesian Inference I

Parametric Inference Maximum Likelihood Inference Exponential Families Expectation Maximization (EM) Bayesian Inference Statistical Decison Theory

Lecture 3. Univariate Bayesian inference: conjugate analysis

The binomial model. Assume a uniform prior distribution on p(θ). Write the pdf for this distribution.

Introduction to Bayesian Inference

Statistical Inference

The Jeffreys Prior. Yingbo Li MATH Clemson University. Yingbo Li (Clemson) The Jeffreys Prior MATH / 13

Lecture 2. (See Exercise 7.22, 7.23, 7.24 in Casella & Berger)

Introduction to Bayesian Methods

Bayesian linear regression

[Chapter 6. Functions of Random Variables]

Modern Methods of Statistical Learning sf2935 Auxiliary material: Exponential Family of Distributions Timo Koski. Second Quarter 2016

Hierarchical models. Dr. Jarad Niemi. August 31, Iowa State University. Jarad Niemi (Iowa State) Hierarchical models August 31, / 31

Part 2: One-parameter models

Bayesian Inference and Decision Theory

DIFFERENT APPROACHES TO STATISTICAL INFERENCE: HYPOTHESIS TESTING VERSUS BAYESIAN ANALYSIS

A Brief Introduction to R Shiny

Bayesian Inference. Chapter 4: Regression and Hierarchical Models

HPD Intervals / Regions

Estimation of Quantiles

INTRODUCTION TO BAYESIAN METHODS II

Statistical Theory MT 2006 Problems 4: Solution sketches

A few basics of credibility theory

10. Exchangeability and hierarchical models Objective. Recommended reading

Statistical Inference

2. Continuous Distributions

Lecture 2: Priors and Conjugacy

ST5215: Advanced Statistical Theory

Bayesian statistics. DS GA 1002 Statistical and Mathematical Models. Carlos Fernandez-Granda

Statistics for scientists and engineers


Other Noninformative Priors

Conjugate Priors: Beta and Normal Spring 2018

Bayesian Inference. Chapter 4: Regression and Hierarchical Models

Time Series and Dynamic Models

Bayesian Methods. David S. Rosenberg. New York University. March 20, 2018

1 Maximum Likelihood Estimation

Hyperparameter estimation in Dirichlet process mixture models

The exponential distribution and the Poisson process

IE 303 Discrete-Event Simulation

Checking for Prior-Data Conflict

Learning Bayesian network : Given structure and completely observed data

Compute f(x θ)f(θ) dθ

Lecture 4. Continuous Random Variables and Transformations of Random Variables

Midterm Examination. STA 215: Statistical Inference. Due Wednesday, 2006 Mar 8, 1:15 pm

Bayesian Point Process Modeling for Extreme Value Analysis, with an Application to Systemic Risk Assessment in Correlated Financial Markets

Graphical Models for Collaborative Filtering


Generalized Bayesian Inference with Sets of Conjugate Priors for Dealing with Prior-Data Conflict

Transcription:

Classical and Bayesian inference AMS 132 January 18, 2018 Claudia Wehrhahn (UCSC) Classical and Bayesian inference January 18, 2018 1 / 9

Sampling from a Bernoulli Distribution Theorem (Beta-Bernoulli model) Suppose that X 1,..., X n form a random sample from the Bernoulli distribution with parameter θ, which is unknown (0 < θ < 1). Suppose also that the prior distribution of θ is the beta distribution with parameters α > 0 and β > 0. Then the posterior distribution of θ given that X i = x i, (i = 1,..., n), is the beta distribution with parameters α + n i=1 x i and β + n n i=1 x i. Example (Beta-Bernoulli model) Consider a large shipment of iphone has arrive, and the proportion of defective iphone, θ, is unknown. A Beta(α, β) prior is assumed and n randomly selected items are inspected. a) Find the posterior distribution of θ. b) How is the posterior distribution updated? Claudia Wehrhahn (UCSC) Classical and Bayesian inference January 18, 2018 2 / 9

Sampling from a Bernoulli Distribution Definition (Conjugate Family and Hyperparameters) Let X 1, X 2,... be conditionally i.i.d. given θ with common p.f. or p.d.f. f (x θ). Let Ψ be a family of possible distributions over the parameter space Ω. Suppose that, no matter which prior distribution ξ we choose from Ψ, no matter how many observations X = (X 1,..., X n) we observe, and no matter what are their observed values x = (x 1,..., x n), the posterior distribution ξ(θ x) is a member of Ψ. Then Ψ is called a conjugate family of prior distributions for samples from the distributions f (x θ). It is also said that the family Ψ is closed under sampling from the distributions f (x θ). Finally, if the distributions in Ψ are parametrized by further parameters, then the associated parameters for the prior distribution are called the prior hyperparameters and the associated parameters of the posterior distribution are called the posterior hyperparameters. Claudia Wehrhahn (UCSC) Classical and Bayesian inference January 18, 2018 3 / 9

Sampling from a Poisson Distribution Theorem (Gamma-Poisson model) Suppose that X 1,..., X n form a random sample from the Poisson distribution with mean θ > 0, which is unknown. Suppose also that the prior distribution of θ is the gamma distribution with parameters α > 0 and β > 0. Then the posterior distribution of θ given that X i = x i, (i = 1,..., n), is the gamma distribution with parameters α + n i=1 x i and β + n. Example (Gamma-Poisson model) Suppose that customers arrive to a store at an unknown rate, θ, per hour. Assume that θ follows a gamma distribution with parameters α and β. a) Find the posterior distribution of θ. Claudia Wehrhahn (UCSC) Classical and Bayesian inference January 18, 2018 4 / 9

Sampling from a Normal Distribution Theorem (Normal-Normal model) Suppose that X 1,..., X n form a random sample from the Normal distribution with unknown mean θ and known variance σ 2 > 0. Suppose also that the prior distribution of θ is the normal distribution with mean µ 0 and variance v 2 0. Then the posterior distribution of θ given that X i = x i, (i = 1,..., n), is the normal distribution with mean and variance given by µ 1 = σ2 µ 0 + nv0 2 x n, v 2 σ 2 + nv0 2 1 = σ2 v0 2. σ 2 + nv0 2 Claudia Wehrhahn (UCSC) Classical and Bayesian inference January 18, 2018 5 / 9

Sampling from a Normal Distribution Example (Normal-Normal model) Suppose that X i θ, σ 2 i.i.d. N(θ, σ 2 ), with σ 2 known. Assume that θ follows a normal distribution with parameters µ 0 and v 2 0. a) Find the posterior distribution of θ. b) Show that the posterior mean can be written as σ 2 µ 1 = σ 2 + nv0 2 c) Find an expression for computing P(θ > 1 x). µ 0 + nv 0 2 x n. σ 2 + nv0 2 Claudia Wehrhahn (UCSC) Classical and Bayesian inference January 18, 2018 6 / 9

Sampling from a Exponential Distribution Theorem (Gamma-Exponential model) Suppose that X 1,..., X n form a random sample from the exponential distribution with unknown parameter θ > 0. Suppose also that the prior distribution of θ is the gamma distribution with parameters α > 0 and β > 0. Then the posterior distribution of θ given that X i = x i, (i = 1,..., n), is the gamma distribution with parameters α + n and β + n i=1 x i. Example (Gamma-Exponential model) Suppose that X i θ i.i.d. exp(θ), i = 1,..., n, θ > 0, where X i describes the lifetime of component i, and θ Gamma(α, β). a) Find the posterior distribution of θ. Claudia Wehrhahn (UCSC) Classical and Bayesian inference January 18, 2018 7 / 9

Improper Prior Distributions Recall that improper priors capture the idea that there is much more information in the data than is capture in our prior distribution. Each of the conjugate families that we have seen has an improper prior as a limiting case. Definition Let ξ be a nonnegative function whose domain includes the parameter space of a statistical model. Suppose that ξ(θ)dθ =. If we pretend as if ξ(θ) is the prior Ω p.d.f. of θ, then we are using an improper prior for θ. Claudia Wehrhahn (UCSC) Classical and Bayesian inference January 18, 2018 8 / 9

Improper Prior Distributions Example (Improper Prior Distributions) a) Consider a large shipment of iphone has arrive, and the proportion of defective iphone, θ, is unknown. A Beta(α, β) prior is assumed and n randomly selected items are inspected. What happens when α = β = 0? b) Suppose that customers arrive to a store at an unknown rate, θ, per hour. Assume that θ follows a gamma distribution with parameters α and β. What happens when α = β = 0? What happens if no customers arrive in an hour? c) Suppose that X i θ, σ 2 i.i.d. N(θ, σ 2 ), with σ 2 known. Assume that θ follows a normal distribution with parameters µ 0 and v 2 0. What happens if v 2 0 =? Claudia Wehrhahn (UCSC) Classical and Bayesian inference January 18, 2018 9 / 9