A novel algorithmic approach to Bayesian Logic Regression

Size: px
Start display at page:

Download "A novel algorithmic approach to Bayesian Logic Regression"

Transcription

1 A novel algorithmic approach to Bayesian Logic Regression Hubin A.A.,Storvik G.O., Frommlet F. Department of Mathematics, University of Oslo and Department of Medical Statistics (CEMSIIS) Oslo Aliaksandr Hubin (University of Oslo) Bayesian Logic Regression / 32

2 Introduction. Issues Logic regression was developed as a tool to construct predictors from Boolean combinations of binary covariates previously used for Inference Not for predictions (before) Among the main applications were Modeling epistatic effects in genetic association studies Regulatory motif finding Identifying target populations for screening or not screening Aliaksandr Hubin (University of Oslo) Bayesian Logic Regression / 32

3 Introduction. Issues Has not become widely known because of Combinatorial complexity Fit algorithms were not performing sufficiently well Few applications were addressed Efficient fit algorithms for model probabilities in space of logic regression are required, since The number of models to select from is doubly exponential in the number of input boolean variables (leaves) The search space has numerous sparsely located local extrema Time and computing resources are limited Aliaksandr Hubin (University of Oslo) Bayesian Logic Regression / 32

4 Bayesian Logic Regression (GLM context) Y i µ i f(y µ i ), i {1,..., n} (1) µ i = g 1 (η i ) (2) η i = γ 0 β 0 + k j=1 γ jβ j L ij (3) L ij {0, 1}, j {1,..., k} are all feasible logical expressions (trees), based on the input leaves. E.g. L i1 = (X i1 X i2 ) Xi3 c, where is logical and is logical or c is logical not k is the total number of all possible trees of size up to C based on p input leaves β j R, j {0,..., k} are regression coefficients of these trees g( ) is a proper link function γ j {0, 1}, j {0,..., k} are latent indicators defining if a tree L ij is included into the model (γ j = 1) or not (γ j = 0) Aliaksandr Hubin (University of Oslo) Bayesian Logic Regression / 32

5 Model Priors p(γ) I { k j=1 γ j Q} k j=1 v γ j c(l j ), (4) c(l j ) is a measure for the complexity of term L j, Q is the maximal allowed number of trees per model, and 0 < v < 1. Aliaksandr Hubin (University of Oslo) Bayesian Logic Regression / 32

6 Model Priors p(γ) I { k j=1 γ j Q} k j=1 v γ j c(l j ), (4) c(l j ) is a measure for the complexity of term L j, Q is the maximal allowed number of trees per model, and 0 < v < 1. Inference driven choice capable of controlling FDR is: p(γ) I { k j=1 γ j Q} k j=1 s j! p sj 2 2s j 2 I {s j C}, (5) s j is a number of leaves in tree L j, p is the number of input leaves, and C is the maximal allowed number of leaves per tree. Aliaksandr Hubin (University of Oslo) Bayesian Logic Regression / 32

7 Model Priors p(γ) I { k j=1 γ j Q} k j=1 v γ j c(l j ), (4) c(l j ) is a measure for the complexity of term L j, Q is the maximal allowed number of trees per model, and 0 < v < 1. Inference driven choice capable of controlling FDR is: p(γ) I { k j=1 γ j Q} k j=1 s j! p sj 2 2s j 2 I {s j C}, (5) s j is a number of leaves in tree L j, p is the number of input leaves, and C is the maximal allowed number of leaves per tree. Prediction or inference driven prior is: p(γ) I {r Q} (1 1/r)(1/r)a (1 1/s)(1/s) b k j=1 I {s j C}, a, b > 0, (6) r = k j=1 γ j is the model size and s in r plus the total number of logical operators present in the model. Aliaksandr Hubin (University of Oslo) Bayesian Logic Regression / 32

8 Marginal likelihood approximation Assume the following prior on regression coefficients: β γ N p (µ β, Σ β ). (7) Then Laplace approximations of the marginal likelihood can be obtained in the GLM context: p(d γ) e log p(d γ,ˆθ γ) 0.5 θ γ log n, (8) where p(d γ, ˆθ γ ) is the likelihood evaluated at the maximum likelihood estimate ˆθ γ of the parameters for model γ (the corresponding regression coefficients and possibly a variance parameter) while n is the number of observations. Aliaksandr Hubin (University of Oslo) Bayesian Logic Regression / 32

9 Inference on the model Let: γ = {γ 1,..., γ k } define a model itself, i.e. which covariates are addressed; θ γ define parameters of the model. Goals: p(γ, θ γ D) posterior distribution of parameters and models; p(γ D) marginal posterior probabilities of the models; p( D) marginal posterior probabilities of the quantiles of interest. But: 2 k different models in Ω γ ; k is huge; Both k and Ω γ are extremely difficult to specify. Aliaksandr Hubin (University of Oslo) Bayesian Logic Regression / 32

10 Possible pipeline Notice that p(γ, θ γ D) = p(θ γ γ, D)p(γ D); Here p(d γ) can be obtained by LA or similarly; Notice that p(γ D) = Approximate with p(d γ)p(γ) γ Ωγ p(d γ )p(γ ) ; p(γ D) = p(d γ)p(γ) γ V p(d γ )p(γ ) (9) V is the subspace of Ω γ to be efficiently explored; Near modal values in terms of MLIK prior are particularly important for construction of reasonable V Ω γ, missing them can dramatically influence posterior in the original space Ω γ. Aliaksandr Hubin (University of Oslo) Bayesian Logic Regression / 32

11 MJMCMC is efficient, but... In Hubin and Storvik [6] we suggested efficient mode jumping proposals in the discrete parameter spaces. But Ω γ and k must be clearly specified for MJMCMC. The later is not feasible in logic regression. Figure: Locally optimized with randomization proposals Aliaksandr Hubin (University of Oslo) Bayesian Logic Regression / 32

12 Genetically modified MJMCMC. Idea MJMCMC is embedded in the iterative setting of a genetic algorithm. In each iteration only a given set S of trees (of fixed size d) is considered; Each S then induces a separate search space for MJMCMC or in the language of genetic algorithms S is the population; S dynamically evolves as a Markov chain of populations through {S 0,..., S tmax } to allow MJMCMC explore different reasonable parts of the in-feasibly large total search space; Each S t, t {1,..., t max } is selected from the neighborhood N t 1 of S t 1. N t 1 includes all populations feasible by performing mutation, crossover, reduction and filtration operations to the current S t 1 ; Utilization of the approximation (9) allows us to compute marginal inclusion probabilities ; p(l D) = p(γ D) (10) γ V L T (γ) Aliaksandr Hubin (University of Oslo) Bayesian Logic Regression / 32

13 Genetically modified MJMCMC. Pipeline S 0 is the set of p input binary leaves; S 1 is constructed by: 1 Running MJMCMC for a given number of iterations N init on S 0 ; 2 The first d 1 < d members of population S 1 are then defined by filtration operation, whilst p d 1 filtered leaves from S 0 are kept in F; 3 The remaining d d 1 members of S 1 are obtained by means of the crossover operation applied to S 0 ; All other S t, t {2,..., t max } are constructed by: 1 Running MJMCMC for a given number of iterations N expl on S t 1 ; 2 The first d t d members of population S t are then defined by filtration operation; 3 The remaining d d t members of S t are obtained by means of the crossover, mutation and reduction operations applied to S t 1 and F; Aliaksandr Hubin (University of Oslo) Bayesian Logic Regression / 32

14 Filtration operation. S 0 case p(f 1 D)... p(f p d1 D) p(l 1 1 D) p(l1 2 D)... p(l1 d 1 D) (11) p(l 1 1 D) po s (12) X 1 F 1 X 2 F p d1 X 3 p o s L 1 1 X p L 1 d 1 Figure: Feature filtering. Aliaksandr Hubin (University of Oslo) Bayesian Logic Regression / 32

15 Filtration operation. S t, t {1,..., t max } case p(d 1 t+1 D)... p(d p d t+1 t+1 D) p(l t+1 d 1 +1 D)... p(lt+1 d t+1 D) (13) p(l t+1 d 1 +1 D) pt s (14) L t d 1 +1 D t+1 1 L t d 1 +2 D t+1 d d t+1 L t d 1 +3 p t s L t+1 d 1 +1 L t d L t+1 d t+1 Figure: Feature filtering. Aliaksandr Hubin (University of Oslo) Bayesian Logic Regression / 32

16 Crossover and mutation operations. Parents selection Other d d t+1 members are filled with either crossovers with probability p c or mutations with probability 1 p c. Crossovers inbreed parents from population S t only, mutations allow parents from F. L t 1 Lt 2 Lt 2... L t d 1 +1 Lt d L t d F 1 F 2 F 3.. Ḟd d1 Figure: Class probabilities for selection of parents proportional to current marginal inclusion probabilities Aliaksandr Hubin (University of Oslo) Bayesian Logic Regression / 32

17 Crossover and mutation operations. Inbreeding of parents Within each mutation or crossover is used for inbreeding with probability p and, and - otherwise. The not c operator is applied to the parents with probability p not. L i L j choose c operators L c i L j Figure: Tree engineering step illustration. Aliaksandr Hubin (University of Oslo) Bayesian Logic Regression / 32

18 Reduction operator Reductions are applied for the trees greater than C; Each leave is independently deleted with Bernoulli probability p d ; The survived leaves are sticked together with with probability p and and - otherwise. L i = X 1 X 2 X 3 X 4 X 5 reduce and stick L i = X 1 X 4 Figure: Tree pruning step illustration. Aliaksandr Hubin (University of Oslo) Bayesian Logic Regression / 32

19 Genetically modified MJMCMC. Embarrassingly parallelized 1 Run B GMJMCMC chains in parallel with different seeds on separate CPUs or clusters; 2 Combine all unique models visited by all B chains into V; 3 Compute model posteriors as (9) 4 Compute marginal inclusion probabilities as (10) 5 Compute posteriors of other parameters of interest as p( D) = p( γ, D) p(γ D) (15) γ V Aliaksandr Hubin (University of Oslo) Bayesian Logic Regression / 32

20 Simulation scenarios. Binary responses For this scenario we generated N = 100 datasets with n = 1000 observations and p = 50 binary covariates. The covariates were assumed to be independent and were simulated as X j Bernoulli(0.3) for j {1,..., 50}. Binary responses Bernoulli observations with individual success probability π: Scenario 1: logit(π) = X1 c X 4 + X 8 X 11 + X 5 X 9 Scenario 2: logit(π) = X c 1 X X 8 X X 5 X 9 Scenario 3: logit(π) = X 2 X 9 +9 X 7 X 12 X 20 9 X 4 X 10 X 17 X 30 Aliaksandr Hubin (University of Oslo) Bayesian Logic Regression / 32

21 Binary responses. Results over 100 datasets FBLR MCLR GMJMCMC Scenario 1 X1 c X X 5 X X 11 X Overall Power FP FDR WL Scenario 2 X1 c X X 5 X X 11 X Overall Power FP FDR WL Scenario 3 X 2 X X 7 X 12 X X 4 X 10 X 17 X Overall Power FP FDR WL Aliaksandr Hubin (University of Oslo) Bayesian Logic Regression / 32

22 Simulation scenarios. Continuous responses For this scenario we generated N = 100 datasets with n = 1000 observations and p = 50 binary covariates. The covariates were assumed to be independent and were simulated as X j Bernoulli(0.5) for j {1,..., 50}. Continuous responses Gaussian observations with error variance σ 2 = 1 and individual expectations specified as follows for the different scenarios: Scenario 4: E(Y ) = X 5 X X 8 X X 1 X 4 Scenario 5: E(Y ) = X X 2 X X 7 X 12 X X 4 X 10 X 17 X 30 Scenario 6: E(Y ) = X X X 18 X X 2 X 9 +9 X 12 X 20 X X 1 X 3 X X 4 X 10 X 17 X X 11 X 13 X 19 X 50 Aliaksandr Hubin (University of Oslo) Bayesian Logic Regression / 32

23 Continuous responses. Results over 100 datasets Scenario 4 Scenario 6 X 5 X X X 8 X X X 1 X X 2 X Overall Power 0.99 X 18 X FP 0.01 X 1 X 3 X FDR X 12 X 20 X WL 0.00 X 4 X 10 X 17 X Scenario 5 X 11 X 13 X 19 X (0.93) X Overall Power 0.79 (0.88) X 2 X FP 4.28 (2.05) X 4 X 10 X 17 X FDR 0.38 (0.19) X 7 X 12 X WL 0.03 Overall Power 0.96 FP 0.37 FDR 0.06 WL 0.00 Aliaksandr Hubin (University of Oslo) Bayesian Logic Regression / 32

24 Real data analysis. Arabadopsis thailana QTL study Phenotype: hypocytl length under different light conditions Population Phenotype Chr Marker expression p(l D) > 0.05 EstC Blue Light 4 X EstC Blue Light 5 X EstC Blue Light 2 X EstC Blue Light 4 2 X X EstC Red Light 2 MSAT EstC Red Light 2 PHYB EstC Red Light 2 1 (1-PHYB) X EstC Red Light 2 X EstC Far Red Light 4 MSAT EstC Far Red Light 4 NGA EstC White Light 5 X EstC White Light 1 X Aliaksandr Hubin (University of Oslo) Bayesian Logic Regression / 32

25 Real data analysis. Drosophila Simulans study Phenotype: pc1 of the size and shape of the posterior lobe of the male genital arch marker chromosome marker name posterior mbic m2 X w x m4 X v x m7 2 gl x m9 2 cg m10 2 gpdh x m14 2 mhc 1 x m18 2 sli x m22 2 zip x m23 2 lsp x m26 3 dbi 1 x m29 3 fz 1 x m32 3 rdg x m33 3 ht m35 3 ninae x m37 3 mst x m40 3 hb m41 3 rox x m44 3 jan 1 x m12, m34 2, 3 glt ant x m11, m35 2, 3 ninae ninac Aliaksandr Hubin (University of Oslo) Bayesian Logic Regression / 32

26 Real data analysis. Drosophila Mauritana study Phenotype: pc1 of the size and shape of the posterior lobe of the male genital arch marker chromosome marker name posterior mbic m1 X ewg x m4 X v x m9 2 cg x m11 2 ninac x m15 2 ddc x m18 2 sli x m22 2 zip x m24 3 ve m25 3 acr x m26 3 dbi m28 3 cyc x m29 3 fz m34 3 ant 1 x m37 3 mst x m39 3 tub m40 3 hb x m41 3 rox m44 3 jan x m1, m2 X, X w ewg m2, m36 X, 3 w.fas x m29, m40 3, 3 fz.hb x Aliaksandr Hubin (University of Oslo) Bayesian Logic Regression / 32

27 Real data analysis. Simulans VS Mauritana prediction Short runs of GMJMCMC with 1 thread are addressed; Less conservative prior is used, namely p(γ) (1 1/r)(1/r)log(500) (1 1/s)(1/s) 2 ; Prediction is based on marginalized over all models probabilities, namely Ŷ = I { p(y D) 0.5}, p(y D) = γ V p(y γ, D) p(γ D); Algorithm min.p mean.p max.p min.fn mean.fn max.fn min.fp mean.fp max.fp RIDGE GMJMCMC NAIVEBAYESS lxgboost MJMCMC LASSO LR DEEPNETS txgboost RFOREST KMEANS Table: Comparison of performance (Precision, FDR, FNR) of different algorithms for Drosophila classification Aliaksandr Hubin (University of Oslo) Bayesian Logic Regression / 32

28 Real data analysis. Simulans VS Mauritana study But no nonlinearities were found either in the inferential study Table: Drosophila data: whether simulans or mauritiana. Population Phenotype Chr Marker expression p(l D) > 0.5 Both S or M X run Both S or M 3 ninae Both S or M 2 ddc Aliaksandr Hubin (University of Oslo) Bayesian Logic Regression / 32

29 Real data analysis. Simulated response prediction Simulate responses as: Y = I {logit 1 (0.4 9v eve eip gpdh + 9gl egfr glt 5cg w) 0.5} Now perform predictions: Algorithm min.p mean.p max.p min.fn mean.fn max.fn min.fp mean.fp max.fp txgboost GMJMCMC MJMCMC lxgboost DEEPNETS RFOREST LASSO RIDGE NAIVEBAYESS LR KMEANS Table: Comparison of performance (Precision, FDR, FNR) of different algorithms for Drosophila classification Aliaksandr Hubin (University of Oslo) Bayesian Logic Regression / 32

30 Further (partly current) research Generalization. trees: Allow general feature engineering instead of only logical L i L j choose + g operator g l (L i L j ) Figure: General feature engineering step illustration. Aliaksandr Hubin (University of Oslo) Bayesian Logic Regression / 32

31 Further (partly current) research Generalization will allow (allows) inference like the following: Table: TP, FP, power and FDR under Kepler s 3rd law model ( p(l D) 0.5). Detections 32 threads 16 threads 4 threads 1 thread Models/Last mutation 5000/ / / /10000 (HostStarMassSlrMass PeriodDays 2 ) (HostStarRadiusSlrRad PeriodDays 2 ) (HostStarTempK PeriodDays 2 ) Other Totally: FDR FP Power Power 32 threads threads threads thread where the observations a are semi major axises of the ellipses of orbits. Explanatory variables include TypeFlag, RadiusJpt, PeriodDays, HostStar- MassSlrMass, Eccentricity, PlanetaryMassJpt, HostStarRadiusSlrRad, Host- StarMetallicity, and PlanetaryDensJpt. Aliaksandr Hubin (University of Oslo) Bayesian Logic Regression / 32

32 Concluding remarks We introduced the GMJMCMC algorithm for Bayesian logic regression models capable of estimating posterior model probabilities Bayesian model averaging and selection EMJMCMC R-package is available flexibility in the choice of methods marginal likelihoods model selection criteria extensive parallel computing is available vectorized predictions with NA hadling is incorporated Results showed that GMJMCMC performs well in terms of the search speed and quality addresses a more general class of models than competitors provides nice predictive and inferential performance in the applications Aliaksandr Hubin (University of Oslo) Bayesian Logic Regression / 32

33 References A. Hubin, G.O. Storvik, F. Frommlet A novel algorithmic approach to Bayesian Logic Regression. arxiv: v1, A. Hubin and G.O. Storvik Efficient mode jumping MCMC for Bayesian variable selection in GLMM. arxiv: v3, C. Kooperberg, and I. Ruczinski. Identifying Interacting SNPs Using Monte Carlo Logic Regression. Genetic Epidemiology, 28: , A. Fritsch. A Full Bayesian Version of Logic regression for SNP Data. PhD thesis, Aliaksandr Hubin (University of Oslo) Bayesian Logic Regression / 32

34 The End. Thank you. Aliaksandr Hubin (University of Oslo) Bayesian Logic Regression / 32

Selecting explanatory variables with the modified version of Bayesian Information Criterion

Selecting explanatory variables with the modified version of Bayesian Information Criterion Selecting explanatory variables with the modified version of Bayesian Information Criterion Institute of Mathematics and Computer Science, Wrocław University of Technology, Poland in cooperation with J.K.Ghosh,

More information

Efficient mode jumping MCMC for Bayesian variable selection in GLMM

Efficient mode jumping MCMC for Bayesian variable selection in GLMM Efficient mode jumping MCMC for Bayesian variable selection in GLMM arxiv:1604.06398v3 [stat.co] 12 Apr 2017 Aliaksandr Hubin Department of Mathematics, University of Oslo and Geir Storvik Department of

More information

Estimating the marginal likelihood with Integrated nested Laplace approximation (INLA)

Estimating the marginal likelihood with Integrated nested Laplace approximation (INLA) Estimating the marginal likelihood with Integrated nested Laplace approximation (INLA) arxiv:1611.01450v1 [stat.co] 4 Nov 2016 Aliaksandr Hubin Department of Mathematics, University of Oslo and Geir Storvik

More information

Bayesian Regression Linear and Logistic Regression

Bayesian Regression Linear and Logistic Regression When we want more than point estimates Bayesian Regression Linear and Logistic Regression Nicole Beckage Ordinary Least Squares Regression and Lasso Regression return only point estimates But what if we

More information

STA414/2104 Statistical Methods for Machine Learning II

STA414/2104 Statistical Methods for Machine Learning II STA414/2104 Statistical Methods for Machine Learning II Murat A. Erdogdu & David Duvenaud Department of Computer Science Department of Statistical Sciences Lecture 3 Slide credits: Russ Salakhutdinov Announcements

More information

An Algorithm for Bayesian Variable Selection in High-dimensional Generalized Linear Models

An Algorithm for Bayesian Variable Selection in High-dimensional Generalized Linear Models Proceedings 59th ISI World Statistics Congress, 25-30 August 2013, Hong Kong (Session CPS023) p.3938 An Algorithm for Bayesian Variable Selection in High-dimensional Generalized Linear Models Vitara Pungpapong

More information

Bayesian construction of perceptrons to predict phenotypes from 584K SNP data.

Bayesian construction of perceptrons to predict phenotypes from 584K SNP data. Bayesian construction of perceptrons to predict phenotypes from 584K SNP data. Luc Janss, Bert Kappen Radboud University Nijmegen Medical Centre Donders Institute for Neuroscience Introduction Genetic

More information

Multiple QTL mapping

Multiple QTL mapping Multiple QTL mapping Karl W Broman Department of Biostatistics Johns Hopkins University www.biostat.jhsph.edu/~kbroman [ Teaching Miscellaneous lectures] 1 Why? Reduce residual variation = increased power

More information

BAYESIAN METHODS FOR VARIABLE SELECTION WITH APPLICATIONS TO HIGH-DIMENSIONAL DATA

BAYESIAN METHODS FOR VARIABLE SELECTION WITH APPLICATIONS TO HIGH-DIMENSIONAL DATA BAYESIAN METHODS FOR VARIABLE SELECTION WITH APPLICATIONS TO HIGH-DIMENSIONAL DATA Intro: Course Outline and Brief Intro to Marina Vannucci Rice University, USA PASI-CIMAT 04/28-30/2010 Marina Vannucci

More information

Marginal Screening and Post-Selection Inference

Marginal Screening and Post-Selection Inference Marginal Screening and Post-Selection Inference Ian McKeague August 13, 2017 Ian McKeague (Columbia University) Marginal Screening August 13, 2017 1 / 29 Outline 1 Background on Marginal Screening 2 2

More information

GWAS IV: Bayesian linear (variance component) models

GWAS IV: Bayesian linear (variance component) models GWAS IV: Bayesian linear (variance component) models Dr. Oliver Stegle Christoh Lippert Prof. Dr. Karsten Borgwardt Max-Planck-Institutes Tübingen, Germany Tübingen Summer 2011 Oliver Stegle GWAS IV: Bayesian

More information

Lecture: Gaussian Process Regression. STAT 6474 Instructor: Hongxiao Zhu

Lecture: Gaussian Process Regression. STAT 6474 Instructor: Hongxiao Zhu Lecture: Gaussian Process Regression STAT 6474 Instructor: Hongxiao Zhu Motivation Reference: Marc Deisenroth s tutorial on Robot Learning. 2 Fast Learning for Autonomous Robots with Gaussian Processes

More information

Pattern Recognition and Machine Learning

Pattern Recognition and Machine Learning Christopher M. Bishop Pattern Recognition and Machine Learning ÖSpri inger Contents Preface Mathematical notation Contents vii xi xiii 1 Introduction 1 1.1 Example: Polynomial Curve Fitting 4 1.2 Probability

More information

Meta-Analysis for Diagnostic Test Data: a Bayesian Approach

Meta-Analysis for Diagnostic Test Data: a Bayesian Approach Meta-Analysis for Diagnostic Test Data: a Bayesian Approach Pablo E. Verde Coordination Centre for Clinical Trials Heinrich Heine Universität Düsseldorf Preliminaries: motivations for systematic reviews

More information

State Space and Hidden Markov Models

State Space and Hidden Markov Models State Space and Hidden Markov Models Kunsch H.R. State Space and Hidden Markov Models. ETH- Zurich Zurich; Aliaksandr Hubin Oslo 2014 Contents 1. Introduction 2. Markov Chains 3. Hidden Markov and State

More information

Sparse Linear Models (10/7/13)

Sparse Linear Models (10/7/13) STA56: Probabilistic machine learning Sparse Linear Models (0/7/) Lecturer: Barbara Engelhardt Scribes: Jiaji Huang, Xin Jiang, Albert Oh Sparsity Sparsity has been a hot topic in statistics and machine

More information

Additional Keplerian Signals in the HARPS data for Gliese 667C from a Bayesian re-analysis

Additional Keplerian Signals in the HARPS data for Gliese 667C from a Bayesian re-analysis Additional Keplerian Signals in the HARPS data for Gliese 667C from a Bayesian re-analysis Phil Gregory, Samantha Lawler, Brett Gladman Physics and Astronomy Univ. of British Columbia Abstract A re-analysis

More information

Contents. Part I: Fundamentals of Bayesian Inference 1

Contents. Part I: Fundamentals of Bayesian Inference 1 Contents Preface xiii Part I: Fundamentals of Bayesian Inference 1 1 Probability and inference 3 1.1 The three steps of Bayesian data analysis 3 1.2 General notation for statistical inference 4 1.3 Bayesian

More information

Bayesian Learning. HT2015: SC4 Statistical Data Mining and Machine Learning. Maximum Likelihood Principle. The Bayesian Learning Framework

Bayesian Learning. HT2015: SC4 Statistical Data Mining and Machine Learning. Maximum Likelihood Principle. The Bayesian Learning Framework HT5: SC4 Statistical Data Mining and Machine Learning Dino Sejdinovic Department of Statistics Oxford http://www.stats.ox.ac.uk/~sejdinov/sdmml.html Maximum Likelihood Principle A generative model for

More information

Part 1: Expectation Propagation

Part 1: Expectation Propagation Chalmers Machine Learning Summer School Approximate message passing and biomedicine Part 1: Expectation Propagation Tom Heskes Machine Learning Group, Institute for Computing and Information Sciences Radboud

More information

Partial factor modeling: predictor-dependent shrinkage for linear regression

Partial factor modeling: predictor-dependent shrinkage for linear regression modeling: predictor-dependent shrinkage for linear Richard Hahn, Carlos Carvalho and Sayan Mukherjee JASA 2013 Review by Esther Salazar Duke University December, 2013 Factor framework The factor framework

More information

Bayesian Classification and Regression Trees

Bayesian Classification and Regression Trees Bayesian Classification and Regression Trees James Cussens York Centre for Complex Systems Analysis & Dept of Computer Science University of York, UK 1 Outline Problems for Lessons from Bayesian phylogeny

More information

STA 4273H: Statistical Machine Learning

STA 4273H: Statistical Machine Learning STA 4273H: Statistical Machine Learning Russ Salakhutdinov Department of Computer Science! Department of Statistical Sciences! rsalakhu@cs.toronto.edu! h0p://www.cs.utoronto.ca/~rsalakhu/ Lecture 7 Approximate

More information

Lecture 6: Graphical Models: Learning

Lecture 6: Graphical Models: Learning Lecture 6: Graphical Models: Learning 4F13: Machine Learning Zoubin Ghahramani and Carl Edward Rasmussen Department of Engineering, University of Cambridge February 3rd, 2010 Ghahramani & Rasmussen (CUED)

More information

Outline. 1 Introduction. 2 Problem and solution. 3 Bayesian tracking model of group debris. 4 Simulation results. 5 Conclusions

Outline. 1 Introduction. 2 Problem and solution. 3 Bayesian tracking model of group debris. 4 Simulation results. 5 Conclusions Outline 1 Introduction 2 Problem and solution 3 Bayesian tracking model of group debris 4 Simulation results 5 Conclusions Problem The limited capability of the radar can t always satisfy the detection

More information

Approximate Bayesian Computation

Approximate Bayesian Computation Approximate Bayesian Computation Michael Gutmann https://sites.google.com/site/michaelgutmann University of Helsinki and Aalto University 1st December 2015 Content Two parts: 1. The basics of approximate

More information

Introduction to Machine Learning

Introduction to Machine Learning Introduction to Machine Learning Brown University CSCI 1950-F, Spring 2012 Prof. Erik Sudderth Lecture 25: Markov Chain Monte Carlo (MCMC) Course Review and Advanced Topics Many figures courtesy Kevin

More information

Bayesian Inference of Interactions and Associations

Bayesian Inference of Interactions and Associations Bayesian Inference of Interactions and Associations Jun Liu Department of Statistics Harvard University http://www.fas.harvard.edu/~junliu Based on collaborations with Yu Zhang, Jing Zhang, Yuan Yuan,

More information

[Part 2] Model Development for the Prediction of Survival Times using Longitudinal Measurements

[Part 2] Model Development for the Prediction of Survival Times using Longitudinal Measurements [Part 2] Model Development for the Prediction of Survival Times using Longitudinal Measurements Aasthaa Bansal PhD Pharmaceutical Outcomes Research & Policy Program University of Washington 69 Biomarkers

More information

Beyond MCMC in fitting complex Bayesian models: The INLA method

Beyond MCMC in fitting complex Bayesian models: The INLA method Beyond MCMC in fitting complex Bayesian models: The INLA method Valeska Andreozzi Centre of Statistics and Applications of Lisbon University (valeska.andreozzi at fc.ul.pt) European Congress of Epidemiology

More information

Multilevel Statistical Models: 3 rd edition, 2003 Contents

Multilevel Statistical Models: 3 rd edition, 2003 Contents Multilevel Statistical Models: 3 rd edition, 2003 Contents Preface Acknowledgements Notation Two and three level models. A general classification notation and diagram Glossary Chapter 1 An introduction

More information

Lecture 2: Linear Models. Bruce Walsh lecture notes Seattle SISG -Mixed Model Course version 23 June 2011

Lecture 2: Linear Models. Bruce Walsh lecture notes Seattle SISG -Mixed Model Course version 23 June 2011 Lecture 2: Linear Models Bruce Walsh lecture notes Seattle SISG -Mixed Model Course version 23 June 2011 1 Quick Review of the Major Points The general linear model can be written as y = X! + e y = vector

More information

Bayesian Methods for Machine Learning

Bayesian Methods for Machine Learning Bayesian Methods for Machine Learning CS 584: Big Data Analytics Material adapted from Radford Neal s tutorial (http://ftp.cs.utoronto.ca/pub/radford/bayes-tut.pdf), Zoubin Ghahramni (http://hunch.net/~coms-4771/zoubin_ghahramani_bayesian_learning.pdf),

More information

Part 2: Multivariate fmri analysis using a sparsifying spatio-temporal prior

Part 2: Multivariate fmri analysis using a sparsifying spatio-temporal prior Chalmers Machine Learning Summer School Approximate message passing and biomedicine Part 2: Multivariate fmri analysis using a sparsifying spatio-temporal prior Tom Heskes joint work with Marcel van Gerven

More information

Bayesian Linear Regression

Bayesian Linear Regression Bayesian Linear Regression Sudipto Banerjee 1 Biostatistics, School of Public Health, University of Minnesota, Minneapolis, Minnesota, U.S.A. September 15, 2010 1 Linear regression models: a Bayesian perspective

More information

Density Estimation. Seungjin Choi

Density Estimation. Seungjin Choi Density Estimation Seungjin Choi Department of Computer Science and Engineering Pohang University of Science and Technology 77 Cheongam-ro, Nam-gu, Pohang 37673, Korea seungjin@postech.ac.kr http://mlg.postech.ac.kr/

More information

Machine Learning Summer School

Machine Learning Summer School Machine Learning Summer School Lecture 3: Learning parameters and structure Zoubin Ghahramani zoubin@eng.cam.ac.uk http://learning.eng.cam.ac.uk/zoubin/ Department of Engineering University of Cambridge,

More information

CPSC 540: Machine Learning

CPSC 540: Machine Learning CPSC 540: Machine Learning MCMC and Non-Parametric Bayes Mark Schmidt University of British Columbia Winter 2016 Admin I went through project proposals: Some of you got a message on Piazza. No news is

More information

Non-Parametric Bayes

Non-Parametric Bayes Non-Parametric Bayes Mark Schmidt UBC Machine Learning Reading Group January 2016 Current Hot Topics in Machine Learning Bayesian learning includes: Gaussian processes. Approximate inference. Bayesian

More information

MCMC for big data. Geir Storvik. BigInsight lunch - May Geir Storvik MCMC for big data BigInsight lunch - May / 17

MCMC for big data. Geir Storvik. BigInsight lunch - May Geir Storvik MCMC for big data BigInsight lunch - May / 17 MCMC for big data Geir Storvik BigInsight lunch - May 2 2018 Geir Storvik MCMC for big data BigInsight lunch - May 2 2018 1 / 17 Outline Why ordinary MCMC is not scalable Different approaches for making

More information

Probabilistic Machine Learning

Probabilistic Machine Learning Probabilistic Machine Learning Bayesian Nets, MCMC, and more Marek Petrik 4/18/2017 Based on: P. Murphy, K. (2012). Machine Learning: A Probabilistic Perspective. Chapter 10. Conditional Independence Independent

More information

Logistic Regression. Seungjin Choi

Logistic Regression. Seungjin Choi Logistic Regression Seungjin Choi Department of Computer Science and Engineering Pohang University of Science and Technology 77 Cheongam-ro, Nam-gu, Pohang 37673, Korea seungjin@postech.ac.kr http://mlg.postech.ac.kr/

More information

Estimation in Generalized Linear Models with Heterogeneous Random Effects. Woncheol Jang Johan Lim. May 19, 2004

Estimation in Generalized Linear Models with Heterogeneous Random Effects. Woncheol Jang Johan Lim. May 19, 2004 Estimation in Generalized Linear Models with Heterogeneous Random Effects Woncheol Jang Johan Lim May 19, 2004 Abstract The penalized quasi-likelihood (PQL) approach is the most common estimation procedure

More information

Probabilistic Graphical Models

Probabilistic Graphical Models Probabilistic Graphical Models Brown University CSCI 2950-P, Spring 2013 Prof. Erik Sudderth Lecture 13: Learning in Gaussian Graphical Models, Non-Gaussian Inference, Monte Carlo Methods Some figures

More information

Logistic regression. 11 Nov Logistic regression (EPFL) Applied Statistics 11 Nov / 20

Logistic regression. 11 Nov Logistic regression (EPFL) Applied Statistics 11 Nov / 20 Logistic regression 11 Nov 2010 Logistic regression (EPFL) Applied Statistics 11 Nov 2010 1 / 20 Modeling overview Want to capture important features of the relationship between a (set of) variable(s)

More information

Related Concepts: Lecture 9 SEM, Statistical Modeling, AI, and Data Mining. I. Terminology of SEM

Related Concepts: Lecture 9 SEM, Statistical Modeling, AI, and Data Mining. I. Terminology of SEM Lecture 9 SEM, Statistical Modeling, AI, and Data Mining I. Terminology of SEM Related Concepts: Causal Modeling Path Analysis Structural Equation Modeling Latent variables (Factors measurable, but thru

More information

an introduction to bayesian inference

an introduction to bayesian inference with an application to network analysis http://jakehofman.com january 13, 2010 motivation would like models that: provide predictive and explanatory power are complex enough to describe observed phenomena

More information

Bayesian Inference. Chapter 4: Regression and Hierarchical Models

Bayesian Inference. Chapter 4: Regression and Hierarchical Models Bayesian Inference Chapter 4: Regression and Hierarchical Models Conchi Ausín and Mike Wiper Department of Statistics Universidad Carlos III de Madrid Advanced Statistics and Data Mining Summer School

More information

Community Health Needs Assessment through Spatial Regression Modeling

Community Health Needs Assessment through Spatial Regression Modeling Community Health Needs Assessment through Spatial Regression Modeling Glen D. Johnson, PhD CUNY School of Public Health glen.johnson@lehman.cuny.edu Objectives: Assess community needs with respect to particular

More information

Generative Clustering, Topic Modeling, & Bayesian Inference

Generative Clustering, Topic Modeling, & Bayesian Inference Generative Clustering, Topic Modeling, & Bayesian Inference INFO-4604, Applied Machine Learning University of Colorado Boulder December 12-14, 2017 Prof. Michael Paul Unsupervised Naïve Bayes Last week

More information

Integrated Non-Factorized Variational Inference

Integrated Non-Factorized Variational Inference Integrated Non-Factorized Variational Inference Shaobo Han, Xuejun Liao and Lawrence Carin Duke University February 27, 2014 S. Han et al. Integrated Non-Factorized Variational Inference February 27, 2014

More information

Random Effects Models for Network Data

Random Effects Models for Network Data Random Effects Models for Network Data Peter D. Hoff 1 Working Paper no. 28 Center for Statistics and the Social Sciences University of Washington Seattle, WA 98195-4320 January 14, 2003 1 Department of

More information

STA 4273H: Statistical Machine Learning

STA 4273H: Statistical Machine Learning STA 4273H: Statistical Machine Learning Russ Salakhutdinov Department of Statistics! rsalakhu@utstat.toronto.edu! http://www.utstat.utoronto.ca/~rsalakhu/ Sidney Smith Hall, Room 6002 Lecture 3 Linear

More information

PMR Learning as Inference

PMR Learning as Inference Outline PMR Learning as Inference Probabilistic Modelling and Reasoning Amos Storkey Modelling 2 The Exponential Family 3 Bayesian Sets School of Informatics, University of Edinburgh Amos Storkey PMR Learning

More information

Statistical Inference and Methods

Statistical Inference and Methods Department of Mathematics Imperial College London d.stephens@imperial.ac.uk http://stats.ma.ic.ac.uk/ das01/ 31st January 2006 Part VI Session 6: Filtering and Time to Event Data Session 6: Filtering and

More information

Logistic Regression Review Fall 2012 Recitation. September 25, 2012 TA: Selen Uguroglu

Logistic Regression Review Fall 2012 Recitation. September 25, 2012 TA: Selen Uguroglu Logistic Regression Review 10-601 Fall 2012 Recitation September 25, 2012 TA: Selen Uguroglu!1 Outline Decision Theory Logistic regression Goal Loss function Inference Gradient Descent!2 Training Data

More information

Package LBLGXE. R topics documented: July 20, Type Package

Package LBLGXE. R topics documented: July 20, Type Package Type Package Package LBLGXE July 20, 2015 Title Bayesian Lasso for detecting Rare (or Common) Haplotype Association and their interactions with Environmental Covariates Version 1.2 Date 2015-07-09 Author

More information

(5) Multi-parameter models - Gibbs sampling. ST440/540: Applied Bayesian Analysis

(5) Multi-parameter models - Gibbs sampling. ST440/540: Applied Bayesian Analysis Summarizing a posterior Given the data and prior the posterior is determined Summarizing the posterior gives parameter estimates, intervals, and hypothesis tests Most of these computations are integrals

More information

Bayesian (conditionally) conjugate inference for discrete data models. Jon Forster (University of Southampton)

Bayesian (conditionally) conjugate inference for discrete data models. Jon Forster (University of Southampton) Bayesian (conditionally) conjugate inference for discrete data models Jon Forster (University of Southampton) with Mark Grigsby (Procter and Gamble?) Emily Webb (Institute of Cancer Research) Table 1:

More information

The sbgcop Package. March 9, 2007

The sbgcop Package. March 9, 2007 The sbgcop Package March 9, 2007 Title Semiparametric Bayesian Gaussian copula estimation Version 0.95 Date 2007-03-09 Author Maintainer This package estimates parameters of

More information

Gaussian Processes. Le Song. Machine Learning II: Advanced Topics CSE 8803ML, Spring 2012

Gaussian Processes. Le Song. Machine Learning II: Advanced Topics CSE 8803ML, Spring 2012 Gaussian Processes Le Song Machine Learning II: Advanced Topics CSE 8803ML, Spring 01 Pictorial view of embedding distribution Transform the entire distribution to expected features Feature space Feature

More information

Bayesian Inference and MCMC

Bayesian Inference and MCMC Bayesian Inference and MCMC Aryan Arbabi Partly based on MCMC slides from CSC412 Fall 2018 1 / 18 Bayesian Inference - Motivation Consider we have a data set D = {x 1,..., x n }. E.g each x i can be the

More information

STA 4273H: Sta-s-cal Machine Learning

STA 4273H: Sta-s-cal Machine Learning STA 4273H: Sta-s-cal Machine Learning Russ Salakhutdinov Department of Computer Science! Department of Statistical Sciences! rsalakhu@cs.toronto.edu! h0p://www.cs.utoronto.ca/~rsalakhu/ Lecture 2 In our

More information

Bayesian Networks: Construction, Inference, Learning and Causal Interpretation. Volker Tresp Summer 2016

Bayesian Networks: Construction, Inference, Learning and Causal Interpretation. Volker Tresp Summer 2016 Bayesian Networks: Construction, Inference, Learning and Causal Interpretation Volker Tresp Summer 2016 1 Introduction So far we were mostly concerned with supervised learning: we predicted one or several

More information

Computational Genomics. Systems biology. Putting it together: Data integration using graphical models

Computational Genomics. Systems biology. Putting it together: Data integration using graphical models 02-710 Computational Genomics Systems biology Putting it together: Data integration using graphical models High throughput data So far in this class we discussed several different types of high throughput

More information

STA414/2104. Lecture 11: Gaussian Processes. Department of Statistics

STA414/2104. Lecture 11: Gaussian Processes. Department of Statistics STA414/2104 Lecture 11: Gaussian Processes Department of Statistics www.utstat.utoronto.ca Delivered by Mark Ebden with thanks to Russ Salakhutdinov Outline Gaussian Processes Exam review Course evaluations

More information

Coupled Hidden Markov Models: Computational Challenges

Coupled Hidden Markov Models: Computational Challenges .. Coupled Hidden Markov Models: Computational Challenges Louis J. M. Aslett and Chris C. Holmes i-like Research Group University of Oxford Warwick Algorithms Seminar 7 th March 2014 ... Hidden Markov

More information

Nonparmeteric Bayes & Gaussian Processes. Baback Moghaddam Machine Learning Group

Nonparmeteric Bayes & Gaussian Processes. Baback Moghaddam Machine Learning Group Nonparmeteric Bayes & Gaussian Processes Baback Moghaddam baback@jpl.nasa.gov Machine Learning Group Outline Bayesian Inference Hierarchical Models Model Selection Parametric vs. Nonparametric Gaussian

More information

Penalized Loss functions for Bayesian Model Choice

Penalized Loss functions for Bayesian Model Choice Penalized Loss functions for Bayesian Model Choice Martyn International Agency for Research on Cancer Lyon, France 13 November 2009 The pure approach For a Bayesian purist, all uncertainty is represented

More information

Bayesian Inference. Chapter 4: Regression and Hierarchical Models

Bayesian Inference. Chapter 4: Regression and Hierarchical Models Bayesian Inference Chapter 4: Regression and Hierarchical Models Conchi Ausín and Mike Wiper Department of Statistics Universidad Carlos III de Madrid Master in Business Administration and Quantitative

More information

STA216: Generalized Linear Models. Lecture 1. Review and Introduction

STA216: Generalized Linear Models. Lecture 1. Review and Introduction STA216: Generalized Linear Models Lecture 1. Review and Introduction Let y 1,..., y n denote n independent observations on a response Treat y i as a realization of a random variable Y i In the general

More information

Default Priors and Effcient Posterior Computation in Bayesian

Default Priors and Effcient Posterior Computation in Bayesian Default Priors and Effcient Posterior Computation in Bayesian Factor Analysis January 16, 2010 Presented by Eric Wang, Duke University Background and Motivation A Brief Review of Parameter Expansion Literature

More information

Bayesian GLMs and Metropolis-Hastings Algorithm

Bayesian GLMs and Metropolis-Hastings Algorithm Bayesian GLMs and Metropolis-Hastings Algorithm We have seen that with conjugate or semi-conjugate prior distributions the Gibbs sampler can be used to sample from the posterior distribution. In situations,

More information

Generalized Linear Models

Generalized Linear Models Generalized Linear Models Advanced Methods for Data Analysis (36-402/36-608 Spring 2014 1 Generalized linear models 1.1 Introduction: two regressions So far we ve seen two canonical settings for regression.

More information

A Review of Pseudo-Marginal Markov Chain Monte Carlo

A Review of Pseudo-Marginal Markov Chain Monte Carlo A Review of Pseudo-Marginal Markov Chain Monte Carlo Discussed by: Yizhe Zhang October 21, 2016 Outline 1 Overview 2 Paper review 3 experiment 4 conclusion Motivation & overview Notation: θ denotes the

More information

Riemann Manifold Methods in Bayesian Statistics

Riemann Manifold Methods in Bayesian Statistics Ricardo Ehlers ehlers@icmc.usp.br Applied Maths and Stats University of São Paulo, Brazil Working Group in Statistical Learning University College Dublin September 2015 Bayesian inference is based on Bayes

More information

Bayesian Inference of Multiple Gaussian Graphical Models

Bayesian Inference of Multiple Gaussian Graphical Models Bayesian Inference of Multiple Gaussian Graphical Models Christine Peterson,, Francesco Stingo, and Marina Vannucci February 18, 2014 Abstract In this paper, we propose a Bayesian approach to inference

More information

Bayesian model selection: methodology, computation and applications

Bayesian model selection: methodology, computation and applications Bayesian model selection: methodology, computation and applications David Nott Department of Statistics and Applied Probability National University of Singapore Statistical Genomics Summer School Program

More information

Design of Text Mining Experiments. Matt Taddy, University of Chicago Booth School of Business faculty.chicagobooth.edu/matt.

Design of Text Mining Experiments. Matt Taddy, University of Chicago Booth School of Business faculty.chicagobooth.edu/matt. Design of Text Mining Experiments Matt Taddy, University of Chicago Booth School of Business faculty.chicagobooth.edu/matt.taddy/research Active Learning: a flavor of design of experiments Optimal : consider

More information

Online appendix to On the stability of the excess sensitivity of aggregate consumption growth in the US

Online appendix to On the stability of the excess sensitivity of aggregate consumption growth in the US Online appendix to On the stability of the excess sensitivity of aggregate consumption growth in the US Gerdie Everaert 1, Lorenzo Pozzi 2, and Ruben Schoonackers 3 1 Ghent University & SHERPPA 2 Erasmus

More information

Linear Dynamical Systems

Linear Dynamical Systems Linear Dynamical Systems Sargur N. srihari@cedar.buffalo.edu Machine Learning Course: http://www.cedar.buffalo.edu/~srihari/cse574/index.html Two Models Described by Same Graph Latent variables Observations

More information

Lecture 2: From Linear Regression to Kalman Filter and Beyond

Lecture 2: From Linear Regression to Kalman Filter and Beyond Lecture 2: From Linear Regression to Kalman Filter and Beyond Department of Biomedical Engineering and Computational Science Aalto University January 26, 2012 Contents 1 Batch and Recursive Estimation

More information

Computational statistics

Computational statistics Computational statistics Combinatorial optimization Thierry Denœux February 2017 Thierry Denœux Computational statistics February 2017 1 / 37 Combinatorial optimization Assume we seek the maximum of f

More information

Estimation of Parameters in Random. Effect Models with Incidence Matrix. Uncertainty

Estimation of Parameters in Random. Effect Models with Incidence Matrix. Uncertainty Estimation of Parameters in Random Effect Models with Incidence Matrix Uncertainty Xia Shen 1,2 and Lars Rönnegård 2,3 1 The Linnaeus Centre for Bioinformatics, Uppsala University, Uppsala, Sweden; 2 School

More information

Introduction to Gaussian Processes

Introduction to Gaussian Processes Introduction to Gaussian Processes Iain Murray murray@cs.toronto.edu CSC255, Introduction to Machine Learning, Fall 28 Dept. Computer Science, University of Toronto The problem Learn scalar function of

More information

STA 414/2104, Spring 2014, Practice Problem Set #1

STA 414/2104, Spring 2014, Practice Problem Set #1 STA 44/4, Spring 4, Practice Problem Set # Note: these problems are not for credit, and not to be handed in Question : Consider a classification problem in which there are two real-valued inputs, and,

More information

Or How to select variables Using Bayesian LASSO

Or How to select variables Using Bayesian LASSO Or How to select variables Using Bayesian LASSO x 1 x 2 x 3 x 4 Or How to select variables Using Bayesian LASSO x 1 x 2 x 3 x 4 Or How to select variables Using Bayesian LASSO On Bayesian Variable Selection

More information

Introduction: MLE, MAP, Bayesian reasoning (28/8/13)

Introduction: MLE, MAP, Bayesian reasoning (28/8/13) STA561: Probabilistic machine learning Introduction: MLE, MAP, Bayesian reasoning (28/8/13) Lecturer: Barbara Engelhardt Scribes: K. Ulrich, J. Subramanian, N. Raval, J. O Hollaren 1 Classifiers In this

More information

Sampling bias in logistic models

Sampling bias in logistic models Sampling bias in logistic models Department of Statistics University of Chicago University of Wisconsin Oct 24, 2007 www.stat.uchicago.edu/~pmcc/reports/bias.pdf Outline Conventional regression models

More information

Bayesian Inference in GLMs. Frequentists typically base inferences on MLEs, asymptotic confidence

Bayesian Inference in GLMs. Frequentists typically base inferences on MLEs, asymptotic confidence Bayesian Inference in GLMs Frequentists typically base inferences on MLEs, asymptotic confidence limits, and log-likelihood ratio tests Bayesians base inferences on the posterior distribution of the unknowns

More information

Kneib, Fahrmeir: Supplement to "Structured additive regression for categorical space-time data: A mixed model approach"

Kneib, Fahrmeir: Supplement to Structured additive regression for categorical space-time data: A mixed model approach Kneib, Fahrmeir: Supplement to "Structured additive regression for categorical space-time data: A mixed model approach" Sonderforschungsbereich 386, Paper 43 (25) Online unter: http://epub.ub.uni-muenchen.de/

More information

10708 Graphical Models: Homework 2

10708 Graphical Models: Homework 2 10708 Graphical Models: Homework 2 Due Monday, March 18, beginning of class Feburary 27, 2013 Instructions: There are five questions (one for extra credit) on this assignment. There is a problem involves

More information

A Brief and Friendly Introduction to Mixed-Effects Models in Linguistics

A Brief and Friendly Introduction to Mixed-Effects Models in Linguistics A Brief and Friendly Introduction to Mixed-Effects Models in Linguistics Cluster-specific parameters ( random effects ) Σb Parameters governing inter-cluster variability b1 b2 bm x11 x1n1 x21 x2n2 xm1

More information

Disease mapping with Gaussian processes

Disease mapping with Gaussian processes EUROHEIS2 Kuopio, Finland 17-18 August 2010 Aki Vehtari (former Helsinki University of Technology) Department of Biomedical Engineering and Computational Science (BECS) Acknowledgments Researchers - Jarno

More information

Nonparametric Bayesian Methods (Gaussian Processes)

Nonparametric Bayesian Methods (Gaussian Processes) [70240413 Statistical Machine Learning, Spring, 2015] Nonparametric Bayesian Methods (Gaussian Processes) Jun Zhu dcszj@mail.tsinghua.edu.cn http://bigml.cs.tsinghua.edu.cn/~jun State Key Lab of Intelligent

More information

Fitting Multidimensional Latent Variable Models using an Efficient Laplace Approximation

Fitting Multidimensional Latent Variable Models using an Efficient Laplace Approximation Fitting Multidimensional Latent Variable Models using an Efficient Laplace Approximation Dimitris Rizopoulos Department of Biostatistics, Erasmus University Medical Center, the Netherlands d.rizopoulos@erasmusmc.nl

More information

Modeling Real Estate Data using Quantile Regression

Modeling Real Estate Data using Quantile Regression Modeling Real Estate Data using Semiparametric Quantile Regression Department of Statistics University of Innsbruck September 9th, 2011 Overview 1 Application: 2 3 4 Hedonic regression data for house prices

More information

Bayesian Image Segmentation Using MRF s Combined with Hierarchical Prior Models

Bayesian Image Segmentation Using MRF s Combined with Hierarchical Prior Models Bayesian Image Segmentation Using MRF s Combined with Hierarchical Prior Models Kohta Aoki 1 and Hiroshi Nagahashi 2 1 Interdisciplinary Graduate School of Science and Engineering, Tokyo Institute of Technology

More information

arxiv:astro-ph/ v1 14 Sep 2005

arxiv:astro-ph/ v1 14 Sep 2005 For publication in Bayesian Inference and Maximum Entropy Methods, San Jose 25, K. H. Knuth, A. E. Abbas, R. D. Morris, J. P. Castle (eds.), AIP Conference Proceeding A Bayesian Analysis of Extrasolar

More information

Linear Regression. Aarti Singh. Machine Learning / Sept 27, 2010

Linear Regression. Aarti Singh. Machine Learning / Sept 27, 2010 Linear Regression Aarti Singh Machine Learning 10-701/15-781 Sept 27, 2010 Discrete to Continuous Labels Classification Sports Science News Anemic cell Healthy cell Regression X = Document Y = Topic X

More information