Bayesian Methods for Highly Correlated Data. Exposures: An Application to Disinfection By-products and Spontaneous Abortion

Size: px
Start display at page:

Download "Bayesian Methods for Highly Correlated Data. Exposures: An Application to Disinfection By-products and Spontaneous Abortion"

Transcription

1 Outline Bayesian Methods for Highly Correlated Exposures: An Application to Disinfection By-products and Spontaneous Abortion November 8, 2007

2 Outline Outline 1 Introduction

3 Outline Outline 1 Introduction 2

4 Outline Outline 1 Introduction 2 3

5 Outline Outline 1 Introduction 2 3 4

6 DBPs and SAB Right From the Start Outline 1 Introduction 2 3 4

7 DBPs and SAB Right From the Start Spontaneous Abortion (SAB) Pregnancy loss prior to 20 weeks gestation Very common (> 30% of all pregnancies) Relatively little known about its causes maternal age, smoking, prior pregnancy loss, occupational exposures, caffeine

8 DBPs and SAB Right From the Start Spontaneous Abortion (SAB) Pregnancy loss prior to 20 weeks gestation Very common (> 30% of all pregnancies) Relatively little known about its causes maternal age, smoking, prior pregnancy loss, occupational exposures, caffeine disinfection by-products (DBPs)?

9 DBPs and SAB Right From the Start Disinfection By-Products A vast array of DBPs are formed in the disinfection process We focus on 2 main types: Trihalomethanes (THMs) CHCl 3, CHBr 3, CHCl 2 Br, CHClBr 2

10 DBPs and SAB Right From the Start Disinfection By-Products A vast array of DBPs are formed in the disinfection process We focus on 2 main types: Trihalomethanes (THMs) CHCl 3, CHBr 3, CHCl 2 Br, CHClBr 2 Haloacetic Acids (HAAs) ClAA, Cl 2 AA, Cl 3 AA, BrAA, Br 2 AA, Br 3 AA, BrClAA, Br 2 ClAA, BrCl 2 AA

11 DBPs and SAB Right From the Start DBPs and SABs Early Studies Noted an increased risk of SAB with increased tap-water consumption

12 DBPs and SAB Right From the Start DBPs and SABs Early Studies Noted an increased risk of SAB with increased tap-water consumption More Recent Studies Increased risk of SAB with exposure to THMs Notably, CHBrCl 2 in Waller et al (1998) OR=2.0 (1.2, 3.5)

13 DBPs and SAB Right From the Start Specific Aim To estimate the effect of each of the 13 constituent DBPs (4 THMs and 9 HAAs) on SAB

14 DBPs and SAB Right From the Start Specific Aim To estimate the effect of each of the 13 constituent DBPs (4 THMs and 9 HAAs) on SAB The Problem DBPs are very highly correlated e.g., ρ=0.91 between Cl 2 AA and Cl 3 AA

15 DBPs and SAB Right From the Start RFTS - briefly 2507 enrolled in three metropolitan areas in U.S. Years: Recruitment Prenatal care practices (52%) Health departments (32%) Promotional mailings (3%) Drug stores, referrals, etc. (13%)

16 DBPs and SAB Right From the Start Eligibility criteria 18 years lived in area served by 1 of the water utilities not using assisted reproductive technology positive pregnancy test intended to carry to term intended to remain in area

17 DBPs and SAB Right From the Start Data Collection Baseline Interview demographic information, medical history, other confounders Pregnancy loss self report or chart abstraction DBP concentration Disinfecting utilities Weekly samples at two sites with high DBPs Every other week at thirdy site with low DBPs

18 Outline Introduction Model Model P1 (semi-bayes) Model P2 (fully Bayes) Dirichlet Process Prior Model (SP1) DPP with selection component (SP2) 1 Introduction 2 3 4

19 Preliminary Analysis Introduction Model Model P1 (semi-bayes) Model P2 (fully Bayes) Dirichlet Process Prior Model (SP1) DPP with selection component (SP2) Discrete time hazard model including all 13 DBPs Time to event: gestational weeks until loss DBP concentrations were measured weekly, included as time-varying covariates Allow for non-linear relationships (crudely) by categorizing DBPs (now have 32 coefficients)

20 Hazard Model Introduction Model Model P1 (semi-bayes) Model P2 (fully Bayes) Dirichlet Process Prior Model (SP1) DPP with selection component (SP2) logit{pr(t i = j T i j, )} = α j + γ 1 z 1i + + γ p z pi + Where, β 1 x 1ij + + β 32 x 32ij α j s are week specific intercepts (weeks ) z 1i... z pi are confounders: smoking, alcohol use, ethnicity, maternal age x kij is the concentration of the k th category of DBP for the i th individual in the j th week

21 of frequentist analysis Model Model P1 (semi-bayes) Model P2 (fully Bayes) Dirichlet Process Prior Model (SP1) DPP with selection component (SP2)

22 of frequentist analysis Model Model P1 (semi-bayes) Model P2 (fully Bayes) Dirichlet Process Prior Model (SP1) DPP with selection component (SP2) Several large but imprecise effects are seen 4 of 32 coefficients are statistically significant Imprecision makes us question results Is there a better analytic approach?

23 Other common options Model Model P1 (semi-bayes) Model P2 (fully Bayes) Dirichlet Process Prior Model (SP1) DPP with selection component (SP2) Try all exposure in one model Problem: unreliable estimates

24 Other common options Model Model P1 (semi-bayes) Model P2 (fully Bayes) Dirichlet Process Prior Model (SP1) DPP with selection component (SP2) Try all exposure in one model Problem: unreliable estimates Combine variables in aggregate scores Problem: difficult to interpret, can mask effects

25 Other common options Model Model P1 (semi-bayes) Model P2 (fully Bayes) Dirichlet Process Prior Model (SP1) DPP with selection component (SP2) Try all exposure in one model Problem: unreliable estimates Combine variables in aggregate scores Problem: difficult to interpret, can mask effects Analyze one variable at a time Problem: uncontrolled confounding

26 Alternative Approaches Model Model P1 (semi-bayes) Model P2 (fully Bayes) Dirichlet Process Prior Model (SP1) DPP with selection component (SP2) Bayesian Parametric Models Semi-Bayes (model P1) Fully-Bayes (model P2) Bayesian Semi-Parametric Models Dirichlet process priors (model SP1) Dirichlet process models with selection component (model SP2)

27 Model P1 Introduction Model Model P1 (semi-bayes) Model P2 (fully Bayes) Dirichlet Process Prior Model (SP1) DPP with selection component (SP2) A simple two-level hierarchical model popularized by Greenland Have seen use in nutritional, genetic, occupational, and cancer epidemiology Despite the name, they are Bayesian models. name may refer to asymptotic methods commonly used in fitting semi-bayes models

28 Model Model P1 (semi-bayes) Model P2 (fully Bayes) Dirichlet Process Prior Model (SP1) DPP with selection component (SP2) Model y i N(x i ) β j N(β 0, φ 2 ) Posterior β N(Ê, V ) V = (X X/σ 2 + I/φ 2 ) 1 Ê = V (X y/σ 2 + β 0 /φ 2 ) y i could be dichotomous and we could use a data augmentation scheme to impute a normal continuous latent variable y i

29 Shrinkage Introduction Model Model P1 (semi-bayes) Model P2 (fully Bayes) Dirichlet Process Prior Model (SP1) DPP with selection component (SP2) Bayesian: natural consequence of combining prior with data Frequentist: introduce bias to reduce MSE (biased but more precise) Amount of shrinkage depends on prior variance

30 Shrinkage in model P1 Model Model P1 (semi-bayes) Model P2 (fully Bayes) Dirichlet Process Prior Model (SP1) DPP with selection component (SP2)

31 Problems with model P1 Model Model P1 (semi-bayes) Model P2 (fully Bayes) Dirichlet Process Prior Model (SP1) DPP with selection component (SP2) Hypothesis testing Have been advocated as a way to reduce problems of multiple comparisons Unfortunately, reduction in type I error rate is typically small

32 Type I error rate with model P1 Model Model P1 (semi-bayes) Model P2 (fully Bayes) Dirichlet Process Prior Model (SP1) DPP with selection component (SP2)

33 More Problems with model P1 Model Model P1 (semi-bayes) Model P2 (fully Bayes) Dirichlet Process Prior Model (SP1) DPP with selection component (SP2) Assumes the prior variance, φ 2, is known with certainty constant shrinkage of all coefficients Sensitivity analyses address changes to results with different prior variances Data contain information on prior variance

34 Model P2 Introduction Model Model P1 (semi-bayes) Model P2 (fully Bayes) Dirichlet Process Prior Model (SP1) DPP with selection component (SP2) Places prior distribution on φ 2 reduces dependence on prior variance Model Specification y i N(x i ) β N(β 0, φ 2 ) φ 2 IG(α 1, α 2 ) Could place prior on µ in some instances as well

35 Model Model P1 (semi-bayes) Model P2 (fully Bayes) Dirichlet Process Prior Model (SP1) DPP with selection component (SP2) What s an inverse-gamma distribution?

36 Properties of model P2 Model Model P1 (semi-bayes) Model P2 (fully Bayes) Dirichlet Process Prior Model (SP1) DPP with selection component (SP2) Prior distribution on φ 2 allows it to be updated by the data As variability of estimates from prior mean increases, so does φ 2 As variability of estimates from prior mean decreases, so does φ 2 Adaptive shrinkage of all coefficients

37 Model Model P1 (semi-bayes) Model P2 (fully Bayes) Dirichlet Process Prior Model (SP1) DPP with selection component (SP2) Model Specification y i N(x i ) β N(β 0, φ 2 ) φ 2 IG(α 1, α 2 ) Conditional Posteriors β y, φ 2 N(Ê, V ) φ 2 β IG(α 1 + p/2, α 2 + (β β 0 ) (β β 0 )/2) V = (X X/σ 2 + I/φ 2 ) 1 Ê = V (X y/σ 2 + β 0 /φ 2 )

38 Adaptive Shrinkage of model P2 Model Model P1 (semi-bayes) Model P2 (fully Bayes) Dirichlet Process Prior Model (SP1) DPP with selection component (SP2) Model Prior variance φ 2, Data Shrinkage P1 Fixed Constant Constant P2 Random

39 Adaptive Shrinkage of model P2 Model Model P1 (semi-bayes) Model P2 (fully Bayes) Dirichlet Process Prior Model (SP1) DPP with selection component (SP2) Model Prior variance φ 2, Data Shrinkage P1 Fixed Constant Constant P2 Random

40 The Problem with Model P2 Model Model P1 (semi-bayes) Model P2 (fully Bayes) Dirichlet Process Prior Model (SP1) DPP with selection component (SP2) How sure are we of our parametric specification of the prior? Can we do better by grouping coefficients into clustering and then shrinking the cluster specific coefficients separately? Amount of shrinkage varies by coefficient

41 Clustering Coefficients Model Model P1 (semi-bayes) Model P2 (fully Bayes) Dirichlet Process Prior Model (SP1) DPP with selection component (SP2)

42 DPP - Model SP1 Introduction Model Model P1 (semi-bayes) Model P2 (fully Bayes) Dirichlet Process Prior Model (SP1) DPP with selection component (SP2) Popular Bayesian non-parametric approach Rather than specifying that β j N(µ, φ 2 ), we specify β j D D is an unknown distribution D needs a prior distribution: D DP(λ, D 0 ) D 0 is a base distribution such as N(µ, φ 2 ) λ is a precision parameter. As λ gets large, D converges to D 0 Sample space Ω with support over D 0 Chop Ω into disjoint Borel sets B 1, B 2,..., B J Then, specifying D DP(αD 0 ) prior implies: (D(B 1 ), D(B 2 ),..., D(B J )) Dirichlet(αD 0 (B 1 ), αd 0 (B 2 ),..., αd 0 (B J ))

43 Random Sample from a DP Model Model P1 (semi-bayes) Model P2 (fully Bayes) Dirichlet Process Prior Model (SP1) DPP with selection component (SP2)

44 Dirichlet Process Priors Model Model P1 (semi-bayes) Model P2 (fully Bayes) Dirichlet Process Prior Model (SP1) DPP with selection component (SP2) β j D, D DP(λ, D 0 ), D 0 N(µ, φ 2 ) This prior implies: β j Data λ λ+k 1 D λ+k 1 i j δ β i β j has a probability of being clustered with any other coefficient Number of expected clusters (asymptotically): λlog(1 + n/λ) Larger values of λ indicate more certainty about the distribution of β j (more clusters) Smaller values of λ indicate less certainty about the distribution of β j (fewer clusters)

45 Dirichlet Process Prior Model Model P1 (semi-bayes) Model P2 (fully Bayes) Dirichlet Process Prior Model (SP1) DPP with selection component (SP2)

46 Clustering Introduction Model Model P1 (semi-bayes) Model P2 (fully Bayes) Dirichlet Process Prior Model (SP1) DPP with selection component (SP2) Consider two parameters, β m and β n, that are equal with probability p mn If β m = β n (p mn = 1), we can use both x im and x in to estimate the common parameter β mn We have twice as much data to estimate the parameter of interest More commonly p mn < 1, so β m adds some information when estimating β n (and vice versa) Part of the reason this method performs so well, is that deciding the probability of clustering is very cheap but the payoff is potentially huge (think of df)

47 Posterior computation Model Model P1 (semi-bayes) Model P2 (fully Bayes) Dirichlet Process Prior Model (SP1) DPP with selection component (SP2) y i N(α + Xβ, σ 2 ) β j D; D DP(λD 0 ); D 0 N(µ, φ 2 ) Conditional Posterior β j β (j), y w 0j D j0 + w kj δ βk k j w oj λ N(β j µ, φ 2 )N(y X j β j, σ 2 ); w kj N(y X j β k, σ 2 ) D j0 N(y X j β k, σ 2 )N(β j µ, φ 2 ) y i = y i α x (j) i β (j)

48 Model SP2 Introduction Model Model P1 (semi-bayes) Model P2 (fully Bayes) Dirichlet Process Prior Model (SP1) DPP with selection component (SP2) Minor modification to Dirichlet process prior model We may desire a more parsimonious model If some DBPs have no effect, would prefer to eliminate them from the model forward/backward selection result in inappropriate confidence intervals

49 Model Model P1 (semi-bayes) Model P2 (fully Bayes) Dirichlet Process Prior Model (SP1) DPP with selection component (SP2) We incorporate a selection model in the Dirichlet Process base distribution: D 0 = πδ 0 + (1 π)n(µ, φ 2 ) π is the probability that a coefficient has no effect (1 π) is the probability that it is N(µ, φ 2 )

50 Model Model P1 (semi-bayes) Model P2 (fully Bayes) Dirichlet Process Prior Model (SP1) DPP with selection component (SP2) A coefficient is equal to zero (no effect) with probability π A priori, we expect this to happen (π100%) of the time We place a prior distribution on π to allow the data to guide inference

51 DPP with variable selection Model Model P1 (semi-bayes) Model P2 (fully Bayes) Dirichlet Process Prior Model (SP1) DPP with selection component (SP2)

52 Simulations Hierarchical Models and RFTS Outline 1 Introduction 2 3 4

53 Simulations Hierarchical Models and RFTS Four hierarchical models, how do they compare? The increased complexity of these hierarchical models seems to make sense, but what does it gain us? Simulated datasets of size n=500

54 Simulations Hierarchical Models and RFTS MSE of Hierarchical Models

55 Simulations Hierarchical Models and RFTS Model P1 logit{pr(t i = j t i j, )} = α j + β 1 x 1i + + β k x ki β j N(µ, φ 2 ) Little prior evidence of effect: specify µ = 0 Calculate φ 2 from existing literature Largest observed effect: OR=3.0 φ 2 = (ln(3) ln(1/3))/(2 1.96) =

56 Simulations Hierarchical Models and RFTS Model P1 -

57 Simulations Hierarchical Models and RFTS Model P2 logit{pr(t i = j t i j, )} = α j + β 1 x 1i + + β k x ki β j N(µ, φ 2 ) φ 2 IG(α 1, α 2 ) µ = 0 φ 2 is random. Choose α 1 = 3.39, α 2 = 1.33 E(φ 2 ) = 0.31 (as in model P1) V (φ 2 ) = 0.07 (at the 95 th percentile of φ 2, 95% of β s will fall between OR=6 and OR=1/6... the most extreme results we believe possible)

58 Simulations Hierarchical Models and RFTS Model P2 -

59 Simulations Hierarchical Models and RFTS Model SP1 logit{pr(t i = j t i j, )} = α j + β 1 x 1i + + β k x ki β j D D DP(λ, D 0 ) D 0 N(µ, φ 2 ) λ G(ν 1, ν 2 ) D 0 N(µ, φ 2 ) φ 2 IG(α 1, α 2 ) µ = 0, α 1 = 3.39, α 2 = 1.33 ν 1 = 1, ν 2 = 1, an uninformative prior for λ

60 Simulations Hierarchical Models and RFTS Model SP1 -

61 Simulations Hierarchical Models and RFTS Model SP2 logit{pr(t i = j t i j, )} = α j + β 1 x 1i + + β k x ki β j D D DP(λ, D 0 ) D 0 N(µ, φ 2 ) λ G(ν 1, ν 2 ) D 0 πδ 0 + (1 π)n(µ, φ 2 ) φ 2 IG(α 1, α 2 ) π beta(ω 1, ω 2 ) µ = 0, α 1 = 3.39, α 2 = 1.33, ν 1 = 1, ν 2 = 1 ω 1 = 1.5, ω 2 = 1.5, so E(π) = 0.5 and 95%CI=(0.01, 0.99)

62 Simulations Hierarchical Models and RFTS Model SP2 -

63 Future/Current Research Outline 1 Introduction 2 3 4

64 Future/Current Research Hierarchical Models Semi-Bayes: Assumes β random Fully-Bayes: Assumes φ 2 random Dirichlet Process: Assumes prior distribution is random Dirichlet Process with Selection Component: Assumes prior distribution is random and allows coefficients to cluster at the null Can improve performance (MSE) with increasing complexity

65 Future/Current Research Hierarchical Models Trade off between performance and difficulty Semi and Fully Bayes are easily implemented in Winbugs (exact distribution) Dirichlet process priors require more programming skill, but often have better MSE

66 Future/Current Research DBPs and SAB Model P1 provided the least shrinkage; Dirichlet Process models, the most These results in contrast to previous research (sort of) Very little evidence of an effect of any constituent DBP on SAB

67 Future/Current Research Dimension reduction in genomics Big problem in genotyping research: P >> N Typical frequentist models won t work in this situation Generally rely on FDR approach: tradeoff between per-comparison error rate and family-wise error rate... but you still need a model to generate your p-values We cluster SNP effects to reduce dimension and use a Double Exponential prior in our base distribution

68 Future/Current Research DP for dimension reduction

69 Future/Current Research Non parametric functional data analysis Functional data is common in epidemiology Use DP prior to fit flexible random curves to data Individual curves may be clustered over all or part of the curve to aid in inference and prediction We use hierarchical DP prior to allow global or local clustering of curves

70 Future/Current Research Functional data analysis

71 Future/Current Research The End

Bayesian Methods for Highly Correlated Exposures: an Application to Tap Water Disinfection By-Products and Spontaneous Abortion

Bayesian Methods for Highly Correlated Exposures: an Application to Tap Water Disinfection By-Products and Spontaneous Abortion Bayesian Methods for Highly Correlated Exposures: an Application to Tap Water Disinfection By-Products and Spontaneous Abortion by Richard F. MacLehose A dissertation submitted to the faculty of the University

More information

Bayesian Regression (1/31/13)

Bayesian Regression (1/31/13) STA613/CBB540: Statistical methods in computational biology Bayesian Regression (1/31/13) Lecturer: Barbara Engelhardt Scribe: Amanda Lea 1 Bayesian Paradigm Bayesian methods ask: given that I have observed

More information

Multivariate Survival Analysis

Multivariate Survival Analysis Multivariate Survival Analysis Previously we have assumed that either (X i, δ i ) or (X i, δ i, Z i ), i = 1,..., n, are i.i.d.. This may not always be the case. Multivariate survival data can arise in

More information

PMR Learning as Inference

PMR Learning as Inference Outline PMR Learning as Inference Probabilistic Modelling and Reasoning Amos Storkey Modelling 2 The Exponential Family 3 Bayesian Sets School of Informatics, University of Edinburgh Amos Storkey PMR Learning

More information

Robustness to Parametric Assumptions in Missing Data Models

Robustness to Parametric Assumptions in Missing Data Models Robustness to Parametric Assumptions in Missing Data Models Bryan Graham NYU Keisuke Hirano University of Arizona April 2011 Motivation Motivation We consider the classic missing data problem. In practice

More information

Bayesian Hypothesis Testing in GLMs: One-Sided and Ordered Alternatives. 1(w i = h + 1)β h + ɛ i,

Bayesian Hypothesis Testing in GLMs: One-Sided and Ordered Alternatives. 1(w i = h + 1)β h + ɛ i, Bayesian Hypothesis Testing in GLMs: One-Sided and Ordered Alternatives Often interest may focus on comparing a null hypothesis of no difference between groups to an ordered restricted alternative. For

More information

Non-Parametric Bayes

Non-Parametric Bayes Non-Parametric Bayes Mark Schmidt UBC Machine Learning Reading Group January 2016 Current Hot Topics in Machine Learning Bayesian learning includes: Gaussian processes. Approximate inference. Bayesian

More information

A Bayesian Nonparametric Approach to Monotone Missing Data in Longitudinal Studies with Informative Missingness

A Bayesian Nonparametric Approach to Monotone Missing Data in Longitudinal Studies with Informative Missingness A Bayesian Nonparametric Approach to Monotone Missing Data in Longitudinal Studies with Informative Missingness A. Linero and M. Daniels UF, UT-Austin SRC 2014, Galveston, TX 1 Background 2 Working model

More information

Selection on Observables: Propensity Score Matching.

Selection on Observables: Propensity Score Matching. Selection on Observables: Propensity Score Matching. Department of Economics and Management Irene Brunetti ireneb@ec.unipi.it 24/10/2017 I. Brunetti Labour Economics in an European Perspective 24/10/2017

More information

Plausible Values for Latent Variables Using Mplus

Plausible Values for Latent Variables Using Mplus Plausible Values for Latent Variables Using Mplus Tihomir Asparouhov and Bengt Muthén August 21, 2010 1 1 Introduction Plausible values are imputed values for latent variables. All latent variables can

More information

Bayesian Multivariate Logistic Regression

Bayesian Multivariate Logistic Regression Bayesian Multivariate Logistic Regression Sean M. O Brien and David B. Dunson Biostatistics Branch National Institute of Environmental Health Sciences Research Triangle Park, NC 1 Goals Brief review of

More information

Nonparametric Bayes tensor factorizations for big data

Nonparametric Bayes tensor factorizations for big data Nonparametric Bayes tensor factorizations for big data David Dunson Department of Statistical Science, Duke University Funded from NIH R01-ES017240, R01-ES017436 & DARPA N66001-09-C-2082 Motivation Conditional

More information

Cluster investigations using Disease mapping methods International workshop on Risk Factors for Childhood Leukemia Berlin May

Cluster investigations using Disease mapping methods International workshop on Risk Factors for Childhood Leukemia Berlin May Cluster investigations using Disease mapping methods International workshop on Risk Factors for Childhood Leukemia Berlin May 5-7 2008 Peter Schlattmann Institut für Biometrie und Klinische Epidemiologie

More information

A comparison of fully Bayesian and two-stage imputation strategies for missing covariate data

A comparison of fully Bayesian and two-stage imputation strategies for missing covariate data A comparison of fully Bayesian and two-stage imputation strategies for missing covariate data Alexina Mason, Sylvia Richardson and Nicky Best Department of Epidemiology and Biostatistics, Imperial College

More information

Lecture 7: Interaction Analysis. Summer Institute in Statistical Genetics 2017

Lecture 7: Interaction Analysis. Summer Institute in Statistical Genetics 2017 Lecture 7: Interaction Analysis Timothy Thornton and Michael Wu Summer Institute in Statistical Genetics 2017 1 / 39 Lecture Outline Beyond main SNP effects Introduction to Concept of Statistical Interaction

More information

Bayesian Linear Regression

Bayesian Linear Regression Bayesian Linear Regression Sudipto Banerjee 1 Biostatistics, School of Public Health, University of Minnesota, Minneapolis, Minnesota, U.S.A. September 15, 2010 1 Linear regression models: a Bayesian perspective

More information

Previous lecture. P-value based combination. Fixed vs random effects models. Meta vs. pooled- analysis. New random effects testing.

Previous lecture. P-value based combination. Fixed vs random effects models. Meta vs. pooled- analysis. New random effects testing. Previous lecture P-value based combination. Fixed vs random effects models. Meta vs. pooled- analysis. New random effects testing. Interaction Outline: Definition of interaction Additive versus multiplicative

More information

Lecture Outline. Biost 518 Applied Biostatistics II. Choice of Model for Analysis. Choice of Model. Choice of Model. Lecture 10: Multiple Regression:

Lecture Outline. Biost 518 Applied Biostatistics II. Choice of Model for Analysis. Choice of Model. Choice of Model. Lecture 10: Multiple Regression: Biost 518 Applied Biostatistics II Scott S. Emerson, M.D., Ph.D. Professor of Biostatistics University of Washington Lecture utline Choice of Model Alternative Models Effect of data driven selection of

More information

Bayesian variable selection via. Penalized credible regions. Brian Reich, NCSU. Joint work with. Howard Bondell and Ander Wilson

Bayesian variable selection via. Penalized credible regions. Brian Reich, NCSU. Joint work with. Howard Bondell and Ander Wilson Bayesian variable selection via penalized credible regions Brian Reich, NC State Joint work with Howard Bondell and Ander Wilson Brian Reich, NCSU Penalized credible regions 1 Motivation big p, small n

More information

Machine Learning Linear Classification. Prof. Matteo Matteucci

Machine Learning Linear Classification. Prof. Matteo Matteucci Machine Learning Linear Classification Prof. Matteo Matteucci Recall from the first lecture 2 X R p Regression Y R Continuous Output X R p Y {Ω 0, Ω 1,, Ω K } Classification Discrete Output X R p Y (X)

More information

Model Selection in GLMs. (should be able to implement frequentist GLM analyses!) Today: standard frequentist methods for model selection

Model Selection in GLMs. (should be able to implement frequentist GLM analyses!) Today: standard frequentist methods for model selection Model Selection in GLMs Last class: estimability/identifiability, analysis of deviance, standard errors & confidence intervals (should be able to implement frequentist GLM analyses!) Today: standard frequentist

More information

CS 340 Fall 2007: Homework 3

CS 340 Fall 2007: Homework 3 CS 34 Fall 27: Homework 3 1 Marginal likelihood for the Beta-Bernoulli model We showed that the marginal likelihood is the ratio of the normalizing constants: p(d) = B(α 1 + N 1, α + N ) B(α 1, α ) = Γ(α

More information

WU Weiterbildung. Linear Mixed Models

WU Weiterbildung. Linear Mixed Models Linear Mixed Effects Models WU Weiterbildung SLIDE 1 Outline 1 Estimation: ML vs. REML 2 Special Models On Two Levels Mixed ANOVA Or Random ANOVA Random Intercept Model Random Coefficients Model Intercept-and-Slopes-as-Outcomes

More information

Nuoo-Ting (Jassy) Molitor, Nicky Best, Chris Jackson and Sylvia Richardson Imperial College UK. September 30, 2008

Nuoo-Ting (Jassy) Molitor, Nicky Best, Chris Jackson and Sylvia Richardson Imperial College UK. September 30, 2008 Using Bayesian graphical models to model biases in observational studies and to combine multiple data sources: Application to low birth-weight and water disinfection by-products Nuoo-Ting (Jassy) Molitor,

More information

A Practitioner s Guide to Cluster-Robust Inference

A Practitioner s Guide to Cluster-Robust Inference A Practitioner s Guide to Cluster-Robust Inference A. C. Cameron and D. L. Miller presented by Federico Curci March 4, 2015 Cameron Miller Cluster Clinic II March 4, 2015 1 / 20 In the previous episode

More information

STAT331. Cox s Proportional Hazards Model

STAT331. Cox s Proportional Hazards Model STAT331 Cox s Proportional Hazards Model In this unit we introduce Cox s proportional hazards (Cox s PH) model, give a heuristic development of the partial likelihood function, and discuss adaptations

More information

Predictive Distributions

Predictive Distributions Predictive Distributions October 6, 2010 Hoff Chapter 4 5 October 5, 2010 Prior Predictive Distribution Before we observe the data, what do we expect the distribution of observations to be? p(y i ) = p(y

More information

Correlation and regression

Correlation and regression 1 Correlation and regression Yongjua Laosiritaworn Introductory on Field Epidemiology 6 July 2015, Thailand Data 2 Illustrative data (Doll, 1955) 3 Scatter plot 4 Doll, 1955 5 6 Correlation coefficient,

More information

Preliminary Statistics. Lecture 5: Hypothesis Testing

Preliminary Statistics. Lecture 5: Hypothesis Testing Preliminary Statistics Lecture 5: Hypothesis Testing Rory Macqueen (rm43@soas.ac.uk), September 2015 Outline Elements/Terminology of Hypothesis Testing Types of Errors Procedure of Testing Significance

More information

Bayesian methods for missing data: part 1. Key Concepts. Nicky Best and Alexina Mason. Imperial College London

Bayesian methods for missing data: part 1. Key Concepts. Nicky Best and Alexina Mason. Imperial College London Bayesian methods for missing data: part 1 Key Concepts Nicky Best and Alexina Mason Imperial College London BAYES 2013, May 21-23, Erasmus University Rotterdam Missing Data: Part 1 BAYES2013 1 / 68 Outline

More information

Bayesian Hierarchical Models

Bayesian Hierarchical Models Bayesian Hierarchical Models Gavin Shaddick, Millie Green, Matthew Thomas University of Bath 6 th - 9 th December 2016 1/ 34 APPLICATIONS OF BAYESIAN HIERARCHICAL MODELS 2/ 34 OUTLINE Spatial epidemiology

More information

Biost 518 Applied Biostatistics II. Purpose of Statistics. First Stage of Scientific Investigation. Further Stages of Scientific Investigation

Biost 518 Applied Biostatistics II. Purpose of Statistics. First Stage of Scientific Investigation. Further Stages of Scientific Investigation Biost 58 Applied Biostatistics II Scott S. Emerson, M.D., Ph.D. Professor of Biostatistics University of Washington Lecture 5: Review Purpose of Statistics Statistics is about science (Science in the broadest

More information

Empirical Bayes Moderation of Asymptotically Linear Parameters

Empirical Bayes Moderation of Asymptotically Linear Parameters Empirical Bayes Moderation of Asymptotically Linear Parameters Nima Hejazi Division of Biostatistics University of California, Berkeley stat.berkeley.edu/~nhejazi nimahejazi.org twitter/@nshejazi github/nhejazi

More information

Bayesian Analysis for Natural Language Processing Lecture 2

Bayesian Analysis for Natural Language Processing Lecture 2 Bayesian Analysis for Natural Language Processing Lecture 2 Shay Cohen February 4, 2013 Administrativia The class has a mailing list: coms-e6998-11@cs.columbia.edu Need two volunteers for leading a discussion

More information

. Also, in this case, p i = N1 ) T, (2) where. I γ C N(N 2 2 F + N1 2 Q)

. Also, in this case, p i = N1 ) T, (2) where. I γ C N(N 2 2 F + N1 2 Q) Supplementary information S7 Testing for association at imputed SPs puted SPs Score tests A Score Test needs calculations of the observed data score and information matrix only under the null hypothesis,

More information

Multi-level Models: Idea

Multi-level Models: Idea Review of 140.656 Review Introduction to multi-level models The two-stage normal-normal model Two-stage linear models with random effects Three-stage linear models Two-stage logistic regression with random

More information

9/26/17. Ridge regression. What our model needs to do. Ridge Regression: L2 penalty. Ridge coefficients. Ridge coefficients

9/26/17. Ridge regression. What our model needs to do. Ridge Regression: L2 penalty. Ridge coefficients. Ridge coefficients What our model needs to do regression Usually, we are not just trying to explain observed data We want to uncover meaningful trends And predict future observations Our questions then are Is β" a good estimate

More information

Supplementary Materials for Molecular QTL Discovery Incorporating Genomic Annotations using Bayesian False Discovery Rate Control

Supplementary Materials for Molecular QTL Discovery Incorporating Genomic Annotations using Bayesian False Discovery Rate Control Supplementary Materials for Molecular QTL Discovery Incorporating Genomic Annotations using Bayesian False Discovery Rate Control Xiaoquan Wen Department of Biostatistics, University of Michigan A Model

More information

Bayesian non-parametric model to longitudinally predict churn

Bayesian non-parametric model to longitudinally predict churn Bayesian non-parametric model to longitudinally predict churn Bruno Scarpa Università di Padova Conference of European Statistics Stakeholders Methodologists, Producers and Users of European Statistics

More information

Introduction to Bayesian Statistics with WinBUGS Part 4 Priors and Hierarchical Models

Introduction to Bayesian Statistics with WinBUGS Part 4 Priors and Hierarchical Models Introduction to Bayesian Statistics with WinBUGS Part 4 Priors and Hierarchical Models Matthew S. Johnson New York ASA Chapter Workshop CUNY Graduate Center New York, NY hspace1in December 17, 2009 December

More information

ECE521 week 3: 23/26 January 2017

ECE521 week 3: 23/26 January 2017 ECE521 week 3: 23/26 January 2017 Outline Probabilistic interpretation of linear regression - Maximum likelihood estimation (MLE) - Maximum a posteriori (MAP) estimation Bias-variance trade-off Linear

More information

Bayesian Inference in GLMs. Frequentists typically base inferences on MLEs, asymptotic confidence

Bayesian Inference in GLMs. Frequentists typically base inferences on MLEs, asymptotic confidence Bayesian Inference in GLMs Frequentists typically base inferences on MLEs, asymptotic confidence limits, and log-likelihood ratio tests Bayesians base inferences on the posterior distribution of the unknowns

More information

BIOS 312: Precision of Statistical Inference

BIOS 312: Precision of Statistical Inference and Power/Sample Size and Standard Errors BIOS 312: of Statistical Inference Chris Slaughter Department of Biostatistics, Vanderbilt University School of Medicine January 3, 2013 Outline Overview and Power/Sample

More information

Statistical Inference: Estimation and Confidence Intervals Hypothesis Testing

Statistical Inference: Estimation and Confidence Intervals Hypothesis Testing Statistical Inference: Estimation and Confidence Intervals Hypothesis Testing 1 In most statistics problems, we assume that the data have been generated from some unknown probability distribution. We desire

More information

Bayesian Methods for Machine Learning

Bayesian Methods for Machine Learning Bayesian Methods for Machine Learning CS 584: Big Data Analytics Material adapted from Radford Neal s tutorial (http://ftp.cs.utoronto.ca/pub/radford/bayes-tut.pdf), Zoubin Ghahramni (http://hunch.net/~coms-4771/zoubin_ghahramani_bayesian_learning.pdf),

More information

Computational Systems Biology: Biology X

Computational Systems Biology: Biology X Bud Mishra Room 1002, 715 Broadway, Courant Institute, NYU, New York, USA L#7:(Mar-23-2010) Genome Wide Association Studies 1 The law of causality... is a relic of a bygone age, surviving, like the monarchy,

More information

Fall 2017 STAT 532 Homework Peter Hoff. 1. Let P be a probability measure on a collection of sets A.

Fall 2017 STAT 532 Homework Peter Hoff. 1. Let P be a probability measure on a collection of sets A. 1. Let P be a probability measure on a collection of sets A. (a) For each n N, let H n be a set in A such that H n H n+1. Show that P (H n ) monotonically converges to P ( k=1 H k) as n. (b) For each n

More information

SIGNAL RANKING-BASED COMPARISON OF AUTOMATIC DETECTION METHODS IN PHARMACOVIGILANCE

SIGNAL RANKING-BASED COMPARISON OF AUTOMATIC DETECTION METHODS IN PHARMACOVIGILANCE SIGNAL RANKING-BASED COMPARISON OF AUTOMATIC DETECTION METHODS IN PHARMACOVIGILANCE A HYPOTHESIS TEST APPROACH Ismaïl Ahmed 1,2, Françoise Haramburu 3,4, Annie Fourrier-Réglat 3,4,5, Frantz Thiessard 4,5,6,

More information

Generative Clustering, Topic Modeling, & Bayesian Inference

Generative Clustering, Topic Modeling, & Bayesian Inference Generative Clustering, Topic Modeling, & Bayesian Inference INFO-4604, Applied Machine Learning University of Colorado Boulder December 12-14, 2017 Prof. Michael Paul Unsupervised Naïve Bayes Last week

More information

Nonparametric Bayesian Methods - Lecture I

Nonparametric Bayesian Methods - Lecture I Nonparametric Bayesian Methods - Lecture I Harry van Zanten Korteweg-de Vries Institute for Mathematics CRiSM Masterclass, April 4-6, 2016 Overview of the lectures I Intro to nonparametric Bayesian statistics

More information

Model Selection Tutorial 2: Problems With Using AIC to Select a Subset of Exposures in a Regression Model

Model Selection Tutorial 2: Problems With Using AIC to Select a Subset of Exposures in a Regression Model Model Selection Tutorial 2: Problems With Using AIC to Select a Subset of Exposures in a Regression Model Centre for Molecular, Environmental, Genetic & Analytic (MEGA) Epidemiology School of Population

More information

Lecture 01: Introduction

Lecture 01: Introduction Lecture 01: Introduction Dipankar Bandyopadhyay, Ph.D. BMTRY 711: Analysis of Categorical Data Spring 2011 Division of Biostatistics and Epidemiology Medical University of South Carolina Lecture 01: Introduction

More information

Bayesian linear regression

Bayesian linear regression Bayesian linear regression Linear regression is the basis of most statistical modeling. The model is Y i = X T i β + ε i, where Y i is the continuous response X i = (X i1,..., X ip ) T is the corresponding

More information

Machine Learning Lecture 2

Machine Learning Lecture 2 Machine Perceptual Learning and Sensory Summer Augmented 15 Computing Many slides adapted from B. Schiele Machine Learning Lecture 2 Probability Density Estimation 16.04.2015 Bastian Leibe RWTH Aachen

More information

Multiple linear regression S6

Multiple linear regression S6 Basic medical statistics for clinical and experimental research Multiple linear regression S6 Katarzyna Jóźwiak k.jozwiak@nki.nl November 15, 2017 1/42 Introduction Two main motivations for doing multiple

More information

Local Likelihood Bayesian Cluster Modeling for small area health data. Andrew Lawson Arnold School of Public Health University of South Carolina

Local Likelihood Bayesian Cluster Modeling for small area health data. Andrew Lawson Arnold School of Public Health University of South Carolina Local Likelihood Bayesian Cluster Modeling for small area health data Andrew Lawson Arnold School of Public Health University of South Carolina Local Likelihood Bayesian Cluster Modelling for Small Area

More information

Theory of Maximum Likelihood Estimation. Konstantin Kashin

Theory of Maximum Likelihood Estimation. Konstantin Kashin Gov 2001 Section 5: Theory of Maximum Likelihood Estimation Konstantin Kashin February 28, 2013 Outline Introduction Likelihood Examples of MLE Variance of MLE Asymptotic Properties What is Statistical

More information

Statistics 203: Introduction to Regression and Analysis of Variance Penalized models

Statistics 203: Introduction to Regression and Analysis of Variance Penalized models Statistics 203: Introduction to Regression and Analysis of Variance Penalized models Jonathan Taylor - p. 1/15 Today s class Bias-Variance tradeoff. Penalized regression. Cross-validation. - p. 2/15 Bias-variance

More information

multilevel modeling: concepts, applications and interpretations

multilevel modeling: concepts, applications and interpretations multilevel modeling: concepts, applications and interpretations lynne c. messer 27 october 2010 warning social and reproductive / perinatal epidemiologist concepts why context matters multilevel models

More information

Accounting for Complex Sample Designs via Mixture Models

Accounting for Complex Sample Designs via Mixture Models Accounting for Complex Sample Designs via Finite Normal Mixture Models 1 1 University of Michigan School of Public Health August 2009 Talk Outline 1 2 Accommodating Sampling Weights in Mixture Models 3

More information

Lecture 8: Information Theory and Statistics

Lecture 8: Information Theory and Statistics Lecture 8: Information Theory and Statistics Part II: Hypothesis Testing and I-Hsiang Wang Department of Electrical Engineering National Taiwan University ihwang@ntu.edu.tw December 23, 2015 1 / 50 I-Hsiang

More information

Propensity Score Weighting with Multilevel Data

Propensity Score Weighting with Multilevel Data Propensity Score Weighting with Multilevel Data Fan Li Department of Statistical Science Duke University October 25, 2012 Joint work with Alan Zaslavsky and Mary Beth Landrum Introduction In comparative

More information

Part 4: Multi-parameter and normal models

Part 4: Multi-parameter and normal models Part 4: Multi-parameter and normal models 1 The normal model Perhaps the most useful (or utilized) probability model for data analysis is the normal distribution There are several reasons for this, e.g.,

More information

CONTENTS OF DAY 2. II. Why Random Sampling is Important 10 A myth, an urban legend, and the real reason NOTES FOR SUMMER STATISTICS INSTITUTE COURSE

CONTENTS OF DAY 2. II. Why Random Sampling is Important 10 A myth, an urban legend, and the real reason NOTES FOR SUMMER STATISTICS INSTITUTE COURSE 1 2 CONTENTS OF DAY 2 I. More Precise Definition of Simple Random Sample 3 Connection with independent random variables 4 Problems with small populations 9 II. Why Random Sampling is Important 10 A myth,

More information

A New Bayesian Variable Selection Method: The Bayesian Lasso with Pseudo Variables

A New Bayesian Variable Selection Method: The Bayesian Lasso with Pseudo Variables A New Bayesian Variable Selection Method: The Bayesian Lasso with Pseudo Variables Qi Tang (Joint work with Kam-Wah Tsui and Sijian Wang) Department of Statistics University of Wisconsin-Madison Feb. 8,

More information

9/12/17. Types of learning. Modeling data. Supervised learning: Classification. Supervised learning: Regression. Unsupervised learning: Clustering

9/12/17. Types of learning. Modeling data. Supervised learning: Classification. Supervised learning: Regression. Unsupervised learning: Clustering Types of learning Modeling data Supervised: we know input and targets Goal is to learn a model that, given input data, accurately predicts target data Unsupervised: we know the input only and want to make

More information

Does low participation in cohort studies induce bias? Additional material

Does low participation in cohort studies induce bias? Additional material Does low participation in cohort studies induce bias? Additional material Content: Page 1: A heuristic proof of the formula for the asymptotic standard error Page 2-3: A description of the simulation study

More information

An introduction to biostatistics: part 1

An introduction to biostatistics: part 1 An introduction to biostatistics: part 1 Cavan Reilly September 6, 2017 Table of contents Introduction to data analysis Uncertainty Probability Conditional probability Random variables Discrete random

More information

Lecture 16 : Bayesian analysis of contingency tables. Bayesian linear regression. Jonathan Marchini (University of Oxford) BS2a MT / 15

Lecture 16 : Bayesian analysis of contingency tables. Bayesian linear regression. Jonathan Marchini (University of Oxford) BS2a MT / 15 Lecture 16 : Bayesian analysis of contingency tables. Bayesian linear regression. Jonathan Marchini (University of Oxford) BS2a MT 2013 1 / 15 Contingency table analysis North Carolina State University

More information

Ignoring the matching variables in cohort studies - when is it valid, and why?

Ignoring the matching variables in cohort studies - when is it valid, and why? Ignoring the matching variables in cohort studies - when is it valid, and why? Arvid Sjölander Abstract In observational studies of the effect of an exposure on an outcome, the exposure-outcome association

More information

Empirical Bayes Moderation of Asymptotically Linear Parameters

Empirical Bayes Moderation of Asymptotically Linear Parameters Empirical Bayes Moderation of Asymptotically Linear Parameters Nima Hejazi Division of Biostatistics University of California, Berkeley stat.berkeley.edu/~nhejazi nimahejazi.org twitter/@nshejazi github/nhejazi

More information

Lecture 7 Time-dependent Covariates in Cox Regression

Lecture 7 Time-dependent Covariates in Cox Regression Lecture 7 Time-dependent Covariates in Cox Regression So far, we ve been considering the following Cox PH model: λ(t Z) = λ 0 (t) exp(β Z) = λ 0 (t) exp( β j Z j ) where β j is the parameter for the the

More information

PS 203 Spring 2002 Homework One - Answer Key

PS 203 Spring 2002 Homework One - Answer Key PS 203 Spring 2002 Homework One - Answer Key 1. If you have a home or office computer, download and install WinBUGS. If you don t have your own computer, try running WinBUGS in the Department lab. 2. The

More information

SRMR in Mplus. Tihomir Asparouhov and Bengt Muthén. May 2, 2018

SRMR in Mplus. Tihomir Asparouhov and Bengt Muthén. May 2, 2018 SRMR in Mplus Tihomir Asparouhov and Bengt Muthén May 2, 2018 1 Introduction In this note we describe the Mplus implementation of the SRMR standardized root mean squared residual) fit index for the models

More information

BIOS 2083 Linear Models c Abdus S. Wahed

BIOS 2083 Linear Models c Abdus S. Wahed Chapter 5 206 Chapter 6 General Linear Model: Statistical Inference 6.1 Introduction So far we have discussed formulation of linear models (Chapter 1), estimability of parameters in a linear model (Chapter

More information

Chapter 8: Sampling distributions of estimators Sections

Chapter 8: Sampling distributions of estimators Sections Chapter 8: Sampling distributions of estimators Sections 8.1 Sampling distribution of a statistic 8.2 The Chi-square distributions 8.3 Joint Distribution of the sample mean and sample variance Skip: p.

More information

FREQUENTIST BEHAVIOR OF FORMAL BAYESIAN INFERENCE

FREQUENTIST BEHAVIOR OF FORMAL BAYESIAN INFERENCE FREQUENTIST BEHAVIOR OF FORMAL BAYESIAN INFERENCE Donald A. Pierce Oregon State Univ (Emeritus), RERF Hiroshima (Retired), Oregon Health Sciences Univ (Adjunct) Ruggero Bellio Univ of Udine For Perugia

More information

The Wishart distribution Scaled Wishart. Wishart Priors. Patrick Breheny. March 28. Patrick Breheny BST 701: Bayesian Modeling in Biostatistics 1/11

The Wishart distribution Scaled Wishart. Wishart Priors. Patrick Breheny. March 28. Patrick Breheny BST 701: Bayesian Modeling in Biostatistics 1/11 Wishart Priors Patrick Breheny March 28 Patrick Breheny BST 701: Bayesian Modeling in Biostatistics 1/11 Introduction When more than two coefficients vary, it becomes difficult to directly model each element

More information

Semiparametric Regression

Semiparametric Regression Semiparametric Regression Patrick Breheny October 22 Patrick Breheny Survival Data Analysis (BIOS 7210) 1/23 Introduction Over the past few weeks, we ve introduced a variety of regression models under

More information

Principles of Bayesian Inference

Principles of Bayesian Inference Principles of Bayesian Inference Sudipto Banerjee University of Minnesota July 20th, 2008 1 Bayesian Principles Classical statistics: model parameters are fixed and unknown. A Bayesian thinks of parameters

More information

Motivation Scale Mixutres of Normals Finite Gaussian Mixtures Skew-Normal Models. Mixture Models. Econ 690. Purdue University

Motivation Scale Mixutres of Normals Finite Gaussian Mixtures Skew-Normal Models. Mixture Models. Econ 690. Purdue University Econ 690 Purdue University In virtually all of the previous lectures, our models have made use of normality assumptions. From a computational point of view, the reason for this assumption is clear: combined

More information

Fundamentals to Biostatistics. Prof. Chandan Chakraborty Associate Professor School of Medical Science & Technology IIT Kharagpur

Fundamentals to Biostatistics. Prof. Chandan Chakraborty Associate Professor School of Medical Science & Technology IIT Kharagpur Fundamentals to Biostatistics Prof. Chandan Chakraborty Associate Professor School of Medical Science & Technology IIT Kharagpur Statistics collection, analysis, interpretation of data development of new

More information

Hierarchical Modeling for Univariate Spatial Data

Hierarchical Modeling for Univariate Spatial Data Hierarchical Modeling for Univariate Spatial Data Geography 890, Hierarchical Bayesian Models for Environmental Spatial Data Analysis February 15, 2011 1 Spatial Domain 2 Geography 890 Spatial Domain This

More information

Machine Learning CSE546 Carlos Guestrin University of Washington. September 30, What about continuous variables?

Machine Learning CSE546 Carlos Guestrin University of Washington. September 30, What about continuous variables? Linear Regression Machine Learning CSE546 Carlos Guestrin University of Washington September 30, 2014 1 What about continuous variables? n Billionaire says: If I am measuring a continuous variable, what

More information

Bayesian Inference on Joint Mixture Models for Survival-Longitudinal Data with Multiple Features. Yangxin Huang

Bayesian Inference on Joint Mixture Models for Survival-Longitudinal Data with Multiple Features. Yangxin Huang Bayesian Inference on Joint Mixture Models for Survival-Longitudinal Data with Multiple Features Yangxin Huang Department of Epidemiology and Biostatistics, COPH, USF, Tampa, FL yhuang@health.usf.edu January

More information

Lecture 4 January 23

Lecture 4 January 23 STAT 263/363: Experimental Design Winter 2016/17 Lecture 4 January 23 Lecturer: Art B. Owen Scribe: Zachary del Rosario 4.1 Bandits Bandits are a form of online (adaptive) experiments; i.e. samples are

More information

Subject CS1 Actuarial Statistics 1 Core Principles

Subject CS1 Actuarial Statistics 1 Core Principles Institute of Actuaries of India Subject CS1 Actuarial Statistics 1 Core Principles For 2019 Examinations Aim The aim of the Actuarial Statistics 1 subject is to provide a grounding in mathematical and

More information

Comparison of multiple imputation methods for systematically and sporadically missing multilevel data

Comparison of multiple imputation methods for systematically and sporadically missing multilevel data Comparison of multiple imputation methods for systematically and sporadically missing multilevel data V. Audigier, I. White, S. Jolani, T. Debray, M. Quartagno, J. Carpenter, S. van Buuren, M. Resche-Rigon

More information

Latent Variable Centering of Predictors and Mediators in Multilevel and Time-Series Models

Latent Variable Centering of Predictors and Mediators in Multilevel and Time-Series Models Latent Variable Centering of Predictors and Mediators in Multilevel and Time-Series Models Tihomir Asparouhov and Bengt Muthén August 5, 2018 Abstract We discuss different methods for centering a predictor

More information

Bayesian Linear Models

Bayesian Linear Models Bayesian Linear Models Sudipto Banerjee September 03 05, 2017 Department of Biostatistics, Fielding School of Public Health, University of California, Los Angeles Linear Regression Linear regression is,

More information

Applied Statistics and Econometrics

Applied Statistics and Econometrics Applied Statistics and Econometrics Lecture 6 Saul Lach September 2017 Saul Lach () Applied Statistics and Econometrics September 2017 1 / 53 Outline of Lecture 6 1 Omitted variable bias (SW 6.1) 2 Multiple

More information

AIM HIGH SCHOOL. Curriculum Map W. 12 Mile Road Farmington Hills, MI (248)

AIM HIGH SCHOOL. Curriculum Map W. 12 Mile Road Farmington Hills, MI (248) AIM HIGH SCHOOL Curriculum Map 2923 W. 12 Mile Road Farmington Hills, MI 48334 (248) 702-6922 www.aimhighschool.com COURSE TITLE: Statistics DESCRIPTION OF COURSE: PREREQUISITES: Algebra 2 Students will

More information

Bayesian Estimation of Regression Coefficients Under Extended Balanced Loss Function

Bayesian Estimation of Regression Coefficients Under Extended Balanced Loss Function Communications in Statistics Theory and Methods, 43: 4253 4264, 2014 Copyright Taylor & Francis Group, LLC ISSN: 0361-0926 print / 1532-415X online DOI: 10.1080/03610926.2012.725498 Bayesian Estimation

More information

Doing Right By Massive Data: How To Bring Probability Modeling To The Analysis Of Huge Datasets Without Taking Over The Datacenter

Doing Right By Massive Data: How To Bring Probability Modeling To The Analysis Of Huge Datasets Without Taking Over The Datacenter Doing Right By Massive Data: How To Bring Probability Modeling To The Analysis Of Huge Datasets Without Taking Over The Datacenter Alexander W Blocker Pavlos Protopapas Xiao-Li Meng 9 February, 2010 Outline

More information

STATS 200: Introduction to Statistical Inference. Lecture 29: Course review

STATS 200: Introduction to Statistical Inference. Lecture 29: Course review STATS 200: Introduction to Statistical Inference Lecture 29: Course review Course review We started in Lecture 1 with a fundamental assumption: Data is a realization of a random process. The goal throughout

More information

Propensity Score Adjustment for Unmeasured Confounding in Observational Studies

Propensity Score Adjustment for Unmeasured Confounding in Observational Studies Propensity Score Adjustment for Unmeasured Confounding in Observational Studies Lawrence C. McCandless Sylvia Richardson Nicky G. Best Department of Epidemiology and Public Health, Imperial College London,

More information

Important note: Transcripts are not substitutes for textbook assignments. 1

Important note: Transcripts are not substitutes for textbook assignments. 1 In this lesson we will cover correlation and regression, two really common statistical analyses for quantitative (or continuous) data. Specially we will review how to organize the data, the importance

More information

Sensitivity Analysis with Several Unmeasured Confounders

Sensitivity Analysis with Several Unmeasured Confounders Sensitivity Analysis with Several Unmeasured Confounders Lawrence McCandless lmccandl@sfu.ca Faculty of Health Sciences, Simon Fraser University, Vancouver Canada Spring 2015 Outline The problem of several

More information

Bayesian Linear Models

Bayesian Linear Models Bayesian Linear Models Sudipto Banerjee 1 and Andrew O. Finley 2 1 Department of Forestry & Department of Geography, Michigan State University, Lansing Michigan, U.S.A. 2 Biostatistics, School of Public

More information

Contents. Part I: Fundamentals of Bayesian Inference 1

Contents. Part I: Fundamentals of Bayesian Inference 1 Contents Preface xiii Part I: Fundamentals of Bayesian Inference 1 1 Probability and inference 3 1.1 The three steps of Bayesian data analysis 3 1.2 General notation for statistical inference 4 1.3 Bayesian

More information