A novel algorithmic approach to Bayesian Logic Regression
|
|
- Candice Newton
- 5 years ago
- Views:
Transcription
1 A novel algorithmic approach to Bayesian Logic Regression Hubin A.A.,Storvik G.O., Frommlet F. Department of Mathematics, University of Oslo and Department of Medical Statistics (CEMSIIS) Oslo Aliaksandr Hubin (University of Oslo) Bayesian Logic Regression / 32
2 Introduction. Issues Logic regression was developed as a tool to construct predictors from Boolean combinations of binary covariates previously used for Inference Not for predictions (before) Among the main applications were Modeling epistatic effects in genetic association studies Regulatory motif finding Identifying target populations for screening or not screening Aliaksandr Hubin (University of Oslo) Bayesian Logic Regression / 32
3 Introduction. Issues Has not become widely known because of Combinatorial complexity Fit algorithms were not performing sufficiently well Few applications were addressed Efficient fit algorithms for model probabilities in space of logic regression are required, since The number of models to select from is doubly exponential in the number of input boolean variables (leaves) The search space has numerous sparsely located local extrema Time and computing resources are limited Aliaksandr Hubin (University of Oslo) Bayesian Logic Regression / 32
4 Bayesian Logic Regression (GLM context) Y i µ i f(y µ i ), i {1,..., n} (1) µ i = g 1 (η i ) (2) η i = γ 0 β 0 + k j=1 γ jβ j L ij (3) L ij {0, 1}, j {1,..., k} are all feasible logical expressions (trees), based on the input leaves. E.g. L i1 = (X i1 X i2 ) Xi3 c, where is logical and is logical or c is logical not k is the total number of all possible trees of size up to C based on p input leaves β j R, j {0,..., k} are regression coefficients of these trees g( ) is a proper link function γ j {0, 1}, j {0,..., k} are latent indicators defining if a tree L ij is included into the model (γ j = 1) or not (γ j = 0) Aliaksandr Hubin (University of Oslo) Bayesian Logic Regression / 32
5 Model Priors p(γ) I { k j=1 γ j Q} k j=1 v γ j c(l j ), (4) c(l j ) is a measure for the complexity of term L j, Q is the maximal allowed number of trees per model, and 0 < v < 1. Aliaksandr Hubin (University of Oslo) Bayesian Logic Regression / 32
6 Model Priors p(γ) I { k j=1 γ j Q} k j=1 v γ j c(l j ), (4) c(l j ) is a measure for the complexity of term L j, Q is the maximal allowed number of trees per model, and 0 < v < 1. Inference driven choice capable of controlling FDR is: p(γ) I { k j=1 γ j Q} k j=1 s j! p sj 2 2s j 2 I {s j C}, (5) s j is a number of leaves in tree L j, p is the number of input leaves, and C is the maximal allowed number of leaves per tree. Aliaksandr Hubin (University of Oslo) Bayesian Logic Regression / 32
7 Model Priors p(γ) I { k j=1 γ j Q} k j=1 v γ j c(l j ), (4) c(l j ) is a measure for the complexity of term L j, Q is the maximal allowed number of trees per model, and 0 < v < 1. Inference driven choice capable of controlling FDR is: p(γ) I { k j=1 γ j Q} k j=1 s j! p sj 2 2s j 2 I {s j C}, (5) s j is a number of leaves in tree L j, p is the number of input leaves, and C is the maximal allowed number of leaves per tree. Prediction or inference driven prior is: p(γ) I {r Q} (1 1/r)(1/r)a (1 1/s)(1/s) b k j=1 I {s j C}, a, b > 0, (6) r = k j=1 γ j is the model size and s in r plus the total number of logical operators present in the model. Aliaksandr Hubin (University of Oslo) Bayesian Logic Regression / 32
8 Marginal likelihood approximation Assume the following prior on regression coefficients: β γ N p (µ β, Σ β ). (7) Then Laplace approximations of the marginal likelihood can be obtained in the GLM context: p(d γ) e log p(d γ,ˆθ γ) 0.5 θ γ log n, (8) where p(d γ, ˆθ γ ) is the likelihood evaluated at the maximum likelihood estimate ˆθ γ of the parameters for model γ (the corresponding regression coefficients and possibly a variance parameter) while n is the number of observations. Aliaksandr Hubin (University of Oslo) Bayesian Logic Regression / 32
9 Inference on the model Let: γ = {γ 1,..., γ k } define a model itself, i.e. which covariates are addressed; θ γ define parameters of the model. Goals: p(γ, θ γ D) posterior distribution of parameters and models; p(γ D) marginal posterior probabilities of the models; p( D) marginal posterior probabilities of the quantiles of interest. But: 2 k different models in Ω γ ; k is huge; Both k and Ω γ are extremely difficult to specify. Aliaksandr Hubin (University of Oslo) Bayesian Logic Regression / 32
10 Possible pipeline Notice that p(γ, θ γ D) = p(θ γ γ, D)p(γ D); Here p(d γ) can be obtained by LA or similarly; Notice that p(γ D) = Approximate with p(d γ)p(γ) γ Ωγ p(d γ )p(γ ) ; p(γ D) = p(d γ)p(γ) γ V p(d γ )p(γ ) (9) V is the subspace of Ω γ to be efficiently explored; Near modal values in terms of MLIK prior are particularly important for construction of reasonable V Ω γ, missing them can dramatically influence posterior in the original space Ω γ. Aliaksandr Hubin (University of Oslo) Bayesian Logic Regression / 32
11 MJMCMC is efficient, but... In Hubin and Storvik [6] we suggested efficient mode jumping proposals in the discrete parameter spaces. But Ω γ and k must be clearly specified for MJMCMC. The later is not feasible in logic regression. Figure: Locally optimized with randomization proposals Aliaksandr Hubin (University of Oslo) Bayesian Logic Regression / 32
12 Genetically modified MJMCMC. Idea MJMCMC is embedded in the iterative setting of a genetic algorithm. In each iteration only a given set S of trees (of fixed size d) is considered; Each S then induces a separate search space for MJMCMC or in the language of genetic algorithms S is the population; S dynamically evolves as a Markov chain of populations through {S 0,..., S tmax } to allow MJMCMC explore different reasonable parts of the in-feasibly large total search space; Each S t, t {1,..., t max } is selected from the neighborhood N t 1 of S t 1. N t 1 includes all populations feasible by performing mutation, crossover, reduction and filtration operations to the current S t 1 ; Utilization of the approximation (9) allows us to compute marginal inclusion probabilities ; p(l D) = p(γ D) (10) γ V L T (γ) Aliaksandr Hubin (University of Oslo) Bayesian Logic Regression / 32
13 Genetically modified MJMCMC. Pipeline S 0 is the set of p input binary leaves; S 1 is constructed by: 1 Running MJMCMC for a given number of iterations N init on S 0 ; 2 The first d 1 < d members of population S 1 are then defined by filtration operation, whilst p d 1 filtered leaves from S 0 are kept in F; 3 The remaining d d 1 members of S 1 are obtained by means of the crossover operation applied to S 0 ; All other S t, t {2,..., t max } are constructed by: 1 Running MJMCMC for a given number of iterations N expl on S t 1 ; 2 The first d t d members of population S t are then defined by filtration operation; 3 The remaining d d t members of S t are obtained by means of the crossover, mutation and reduction operations applied to S t 1 and F; Aliaksandr Hubin (University of Oslo) Bayesian Logic Regression / 32
14 Filtration operation. S 0 case p(f 1 D)... p(f p d1 D) p(l 1 1 D) p(l1 2 D)... p(l1 d 1 D) (11) p(l 1 1 D) po s (12) X 1 F 1 X 2 F p d1 X 3 p o s L 1 1 X p L 1 d 1 Figure: Feature filtering. Aliaksandr Hubin (University of Oslo) Bayesian Logic Regression / 32
15 Filtration operation. S t, t {1,..., t max } case p(d 1 t+1 D)... p(d p d t+1 t+1 D) p(l t+1 d 1 +1 D)... p(lt+1 d t+1 D) (13) p(l t+1 d 1 +1 D) pt s (14) L t d 1 +1 D t+1 1 L t d 1 +2 D t+1 d d t+1 L t d 1 +3 p t s L t+1 d 1 +1 L t d L t+1 d t+1 Figure: Feature filtering. Aliaksandr Hubin (University of Oslo) Bayesian Logic Regression / 32
16 Crossover and mutation operations. Parents selection Other d d t+1 members are filled with either crossovers with probability p c or mutations with probability 1 p c. Crossovers inbreed parents from population S t only, mutations allow parents from F. L t 1 Lt 2 Lt 2... L t d 1 +1 Lt d L t d F 1 F 2 F 3.. Ḟd d1 Figure: Class probabilities for selection of parents proportional to current marginal inclusion probabilities Aliaksandr Hubin (University of Oslo) Bayesian Logic Regression / 32
17 Crossover and mutation operations. Inbreeding of parents Within each mutation or crossover is used for inbreeding with probability p and, and - otherwise. The not c operator is applied to the parents with probability p not. L i L j choose c operators L c i L j Figure: Tree engineering step illustration. Aliaksandr Hubin (University of Oslo) Bayesian Logic Regression / 32
18 Reduction operator Reductions are applied for the trees greater than C; Each leave is independently deleted with Bernoulli probability p d ; The survived leaves are sticked together with with probability p and and - otherwise. L i = X 1 X 2 X 3 X 4 X 5 reduce and stick L i = X 1 X 4 Figure: Tree pruning step illustration. Aliaksandr Hubin (University of Oslo) Bayesian Logic Regression / 32
19 Genetically modified MJMCMC. Embarrassingly parallelized 1 Run B GMJMCMC chains in parallel with different seeds on separate CPUs or clusters; 2 Combine all unique models visited by all B chains into V; 3 Compute model posteriors as (9) 4 Compute marginal inclusion probabilities as (10) 5 Compute posteriors of other parameters of interest as p( D) = p( γ, D) p(γ D) (15) γ V Aliaksandr Hubin (University of Oslo) Bayesian Logic Regression / 32
20 Simulation scenarios. Binary responses For this scenario we generated N = 100 datasets with n = 1000 observations and p = 50 binary covariates. The covariates were assumed to be independent and were simulated as X j Bernoulli(0.3) for j {1,..., 50}. Binary responses Bernoulli observations with individual success probability π: Scenario 1: logit(π) = X1 c X 4 + X 8 X 11 + X 5 X 9 Scenario 2: logit(π) = X c 1 X X 8 X X 5 X 9 Scenario 3: logit(π) = X 2 X 9 +9 X 7 X 12 X 20 9 X 4 X 10 X 17 X 30 Aliaksandr Hubin (University of Oslo) Bayesian Logic Regression / 32
21 Binary responses. Results over 100 datasets FBLR MCLR GMJMCMC Scenario 1 X1 c X X 5 X X 11 X Overall Power FP FDR WL Scenario 2 X1 c X X 5 X X 11 X Overall Power FP FDR WL Scenario 3 X 2 X X 7 X 12 X X 4 X 10 X 17 X Overall Power FP FDR WL Aliaksandr Hubin (University of Oslo) Bayesian Logic Regression / 32
22 Simulation scenarios. Continuous responses For this scenario we generated N = 100 datasets with n = 1000 observations and p = 50 binary covariates. The covariates were assumed to be independent and were simulated as X j Bernoulli(0.5) for j {1,..., 50}. Continuous responses Gaussian observations with error variance σ 2 = 1 and individual expectations specified as follows for the different scenarios: Scenario 4: E(Y ) = X 5 X X 8 X X 1 X 4 Scenario 5: E(Y ) = X X 2 X X 7 X 12 X X 4 X 10 X 17 X 30 Scenario 6: E(Y ) = X X X 18 X X 2 X 9 +9 X 12 X 20 X X 1 X 3 X X 4 X 10 X 17 X X 11 X 13 X 19 X 50 Aliaksandr Hubin (University of Oslo) Bayesian Logic Regression / 32
23 Continuous responses. Results over 100 datasets Scenario 4 Scenario 6 X 5 X X X 8 X X X 1 X X 2 X Overall Power 0.99 X 18 X FP 0.01 X 1 X 3 X FDR X 12 X 20 X WL 0.00 X 4 X 10 X 17 X Scenario 5 X 11 X 13 X 19 X (0.93) X Overall Power 0.79 (0.88) X 2 X FP 4.28 (2.05) X 4 X 10 X 17 X FDR 0.38 (0.19) X 7 X 12 X WL 0.03 Overall Power 0.96 FP 0.37 FDR 0.06 WL 0.00 Aliaksandr Hubin (University of Oslo) Bayesian Logic Regression / 32
24 Real data analysis. Arabadopsis thailana QTL study Phenotype: hypocytl length under different light conditions Population Phenotype Chr Marker expression p(l D) > 0.05 EstC Blue Light 4 X EstC Blue Light 5 X EstC Blue Light 2 X EstC Blue Light 4 2 X X EstC Red Light 2 MSAT EstC Red Light 2 PHYB EstC Red Light 2 1 (1-PHYB) X EstC Red Light 2 X EstC Far Red Light 4 MSAT EstC Far Red Light 4 NGA EstC White Light 5 X EstC White Light 1 X Aliaksandr Hubin (University of Oslo) Bayesian Logic Regression / 32
25 Real data analysis. Drosophila Simulans study Phenotype: pc1 of the size and shape of the posterior lobe of the male genital arch marker chromosome marker name posterior mbic m2 X w x m4 X v x m7 2 gl x m9 2 cg m10 2 gpdh x m14 2 mhc 1 x m18 2 sli x m22 2 zip x m23 2 lsp x m26 3 dbi 1 x m29 3 fz 1 x m32 3 rdg x m33 3 ht m35 3 ninae x m37 3 mst x m40 3 hb m41 3 rox x m44 3 jan 1 x m12, m34 2, 3 glt ant x m11, m35 2, 3 ninae ninac Aliaksandr Hubin (University of Oslo) Bayesian Logic Regression / 32
26 Real data analysis. Drosophila Mauritana study Phenotype: pc1 of the size and shape of the posterior lobe of the male genital arch marker chromosome marker name posterior mbic m1 X ewg x m4 X v x m9 2 cg x m11 2 ninac x m15 2 ddc x m18 2 sli x m22 2 zip x m24 3 ve m25 3 acr x m26 3 dbi m28 3 cyc x m29 3 fz m34 3 ant 1 x m37 3 mst x m39 3 tub m40 3 hb x m41 3 rox m44 3 jan x m1, m2 X, X w ewg m2, m36 X, 3 w.fas x m29, m40 3, 3 fz.hb x Aliaksandr Hubin (University of Oslo) Bayesian Logic Regression / 32
27 Real data analysis. Simulans VS Mauritana prediction Short runs of GMJMCMC with 1 thread are addressed; Less conservative prior is used, namely p(γ) (1 1/r)(1/r)log(500) (1 1/s)(1/s) 2 ; Prediction is based on marginalized over all models probabilities, namely Ŷ = I { p(y D) 0.5}, p(y D) = γ V p(y γ, D) p(γ D); Algorithm min.p mean.p max.p min.fn mean.fn max.fn min.fp mean.fp max.fp RIDGE GMJMCMC NAIVEBAYESS lxgboost MJMCMC LASSO LR DEEPNETS txgboost RFOREST KMEANS Table: Comparison of performance (Precision, FDR, FNR) of different algorithms for Drosophila classification Aliaksandr Hubin (University of Oslo) Bayesian Logic Regression / 32
28 Real data analysis. Simulans VS Mauritana study But no nonlinearities were found either in the inferential study Table: Drosophila data: whether simulans or mauritiana. Population Phenotype Chr Marker expression p(l D) > 0.5 Both S or M X run Both S or M 3 ninae Both S or M 2 ddc Aliaksandr Hubin (University of Oslo) Bayesian Logic Regression / 32
29 Real data analysis. Simulated response prediction Simulate responses as: Y = I {logit 1 (0.4 9v eve eip gpdh + 9gl egfr glt 5cg w) 0.5} Now perform predictions: Algorithm min.p mean.p max.p min.fn mean.fn max.fn min.fp mean.fp max.fp txgboost GMJMCMC MJMCMC lxgboost DEEPNETS RFOREST LASSO RIDGE NAIVEBAYESS LR KMEANS Table: Comparison of performance (Precision, FDR, FNR) of different algorithms for Drosophila classification Aliaksandr Hubin (University of Oslo) Bayesian Logic Regression / 32
30 Further (partly current) research Generalization. trees: Allow general feature engineering instead of only logical L i L j choose + g operator g l (L i L j ) Figure: General feature engineering step illustration. Aliaksandr Hubin (University of Oslo) Bayesian Logic Regression / 32
31 Further (partly current) research Generalization will allow (allows) inference like the following: Table: TP, FP, power and FDR under Kepler s 3rd law model ( p(l D) 0.5). Detections 32 threads 16 threads 4 threads 1 thread Models/Last mutation 5000/ / / /10000 (HostStarMassSlrMass PeriodDays 2 ) (HostStarRadiusSlrRad PeriodDays 2 ) (HostStarTempK PeriodDays 2 ) Other Totally: FDR FP Power Power 32 threads threads threads thread where the observations a are semi major axises of the ellipses of orbits. Explanatory variables include TypeFlag, RadiusJpt, PeriodDays, HostStar- MassSlrMass, Eccentricity, PlanetaryMassJpt, HostStarRadiusSlrRad, Host- StarMetallicity, and PlanetaryDensJpt. Aliaksandr Hubin (University of Oslo) Bayesian Logic Regression / 32
32 Concluding remarks We introduced the GMJMCMC algorithm for Bayesian logic regression models capable of estimating posterior model probabilities Bayesian model averaging and selection EMJMCMC R-package is available flexibility in the choice of methods marginal likelihoods model selection criteria extensive parallel computing is available vectorized predictions with NA hadling is incorporated Results showed that GMJMCMC performs well in terms of the search speed and quality addresses a more general class of models than competitors provides nice predictive and inferential performance in the applications Aliaksandr Hubin (University of Oslo) Bayesian Logic Regression / 32
33 References A. Hubin, G.O. Storvik, F. Frommlet A novel algorithmic approach to Bayesian Logic Regression. arxiv: v1, A. Hubin and G.O. Storvik Efficient mode jumping MCMC for Bayesian variable selection in GLMM. arxiv: v3, C. Kooperberg, and I. Ruczinski. Identifying Interacting SNPs Using Monte Carlo Logic Regression. Genetic Epidemiology, 28: , A. Fritsch. A Full Bayesian Version of Logic regression for SNP Data. PhD thesis, Aliaksandr Hubin (University of Oslo) Bayesian Logic Regression / 32
34 The End. Thank you. Aliaksandr Hubin (University of Oslo) Bayesian Logic Regression / 32
Selecting explanatory variables with the modified version of Bayesian Information Criterion
Selecting explanatory variables with the modified version of Bayesian Information Criterion Institute of Mathematics and Computer Science, Wrocław University of Technology, Poland in cooperation with J.K.Ghosh,
More informationEfficient mode jumping MCMC for Bayesian variable selection in GLMM
Efficient mode jumping MCMC for Bayesian variable selection in GLMM arxiv:1604.06398v3 [stat.co] 12 Apr 2017 Aliaksandr Hubin Department of Mathematics, University of Oslo and Geir Storvik Department of
More informationEstimating the marginal likelihood with Integrated nested Laplace approximation (INLA)
Estimating the marginal likelihood with Integrated nested Laplace approximation (INLA) arxiv:1611.01450v1 [stat.co] 4 Nov 2016 Aliaksandr Hubin Department of Mathematics, University of Oslo and Geir Storvik
More informationBayesian Regression Linear and Logistic Regression
When we want more than point estimates Bayesian Regression Linear and Logistic Regression Nicole Beckage Ordinary Least Squares Regression and Lasso Regression return only point estimates But what if we
More informationSTA414/2104 Statistical Methods for Machine Learning II
STA414/2104 Statistical Methods for Machine Learning II Murat A. Erdogdu & David Duvenaud Department of Computer Science Department of Statistical Sciences Lecture 3 Slide credits: Russ Salakhutdinov Announcements
More informationAn Algorithm for Bayesian Variable Selection in High-dimensional Generalized Linear Models
Proceedings 59th ISI World Statistics Congress, 25-30 August 2013, Hong Kong (Session CPS023) p.3938 An Algorithm for Bayesian Variable Selection in High-dimensional Generalized Linear Models Vitara Pungpapong
More informationBayesian construction of perceptrons to predict phenotypes from 584K SNP data.
Bayesian construction of perceptrons to predict phenotypes from 584K SNP data. Luc Janss, Bert Kappen Radboud University Nijmegen Medical Centre Donders Institute for Neuroscience Introduction Genetic
More informationMultiple QTL mapping
Multiple QTL mapping Karl W Broman Department of Biostatistics Johns Hopkins University www.biostat.jhsph.edu/~kbroman [ Teaching Miscellaneous lectures] 1 Why? Reduce residual variation = increased power
More informationBAYESIAN METHODS FOR VARIABLE SELECTION WITH APPLICATIONS TO HIGH-DIMENSIONAL DATA
BAYESIAN METHODS FOR VARIABLE SELECTION WITH APPLICATIONS TO HIGH-DIMENSIONAL DATA Intro: Course Outline and Brief Intro to Marina Vannucci Rice University, USA PASI-CIMAT 04/28-30/2010 Marina Vannucci
More informationMarginal Screening and Post-Selection Inference
Marginal Screening and Post-Selection Inference Ian McKeague August 13, 2017 Ian McKeague (Columbia University) Marginal Screening August 13, 2017 1 / 29 Outline 1 Background on Marginal Screening 2 2
More informationGWAS IV: Bayesian linear (variance component) models
GWAS IV: Bayesian linear (variance component) models Dr. Oliver Stegle Christoh Lippert Prof. Dr. Karsten Borgwardt Max-Planck-Institutes Tübingen, Germany Tübingen Summer 2011 Oliver Stegle GWAS IV: Bayesian
More informationLecture: Gaussian Process Regression. STAT 6474 Instructor: Hongxiao Zhu
Lecture: Gaussian Process Regression STAT 6474 Instructor: Hongxiao Zhu Motivation Reference: Marc Deisenroth s tutorial on Robot Learning. 2 Fast Learning for Autonomous Robots with Gaussian Processes
More informationPattern Recognition and Machine Learning
Christopher M. Bishop Pattern Recognition and Machine Learning ÖSpri inger Contents Preface Mathematical notation Contents vii xi xiii 1 Introduction 1 1.1 Example: Polynomial Curve Fitting 4 1.2 Probability
More informationMeta-Analysis for Diagnostic Test Data: a Bayesian Approach
Meta-Analysis for Diagnostic Test Data: a Bayesian Approach Pablo E. Verde Coordination Centre for Clinical Trials Heinrich Heine Universität Düsseldorf Preliminaries: motivations for systematic reviews
More informationState Space and Hidden Markov Models
State Space and Hidden Markov Models Kunsch H.R. State Space and Hidden Markov Models. ETH- Zurich Zurich; Aliaksandr Hubin Oslo 2014 Contents 1. Introduction 2. Markov Chains 3. Hidden Markov and State
More informationSparse Linear Models (10/7/13)
STA56: Probabilistic machine learning Sparse Linear Models (0/7/) Lecturer: Barbara Engelhardt Scribes: Jiaji Huang, Xin Jiang, Albert Oh Sparsity Sparsity has been a hot topic in statistics and machine
More informationAdditional Keplerian Signals in the HARPS data for Gliese 667C from a Bayesian re-analysis
Additional Keplerian Signals in the HARPS data for Gliese 667C from a Bayesian re-analysis Phil Gregory, Samantha Lawler, Brett Gladman Physics and Astronomy Univ. of British Columbia Abstract A re-analysis
More informationContents. Part I: Fundamentals of Bayesian Inference 1
Contents Preface xiii Part I: Fundamentals of Bayesian Inference 1 1 Probability and inference 3 1.1 The three steps of Bayesian data analysis 3 1.2 General notation for statistical inference 4 1.3 Bayesian
More informationBayesian Learning. HT2015: SC4 Statistical Data Mining and Machine Learning. Maximum Likelihood Principle. The Bayesian Learning Framework
HT5: SC4 Statistical Data Mining and Machine Learning Dino Sejdinovic Department of Statistics Oxford http://www.stats.ox.ac.uk/~sejdinov/sdmml.html Maximum Likelihood Principle A generative model for
More informationPart 1: Expectation Propagation
Chalmers Machine Learning Summer School Approximate message passing and biomedicine Part 1: Expectation Propagation Tom Heskes Machine Learning Group, Institute for Computing and Information Sciences Radboud
More informationPartial factor modeling: predictor-dependent shrinkage for linear regression
modeling: predictor-dependent shrinkage for linear Richard Hahn, Carlos Carvalho and Sayan Mukherjee JASA 2013 Review by Esther Salazar Duke University December, 2013 Factor framework The factor framework
More informationBayesian Classification and Regression Trees
Bayesian Classification and Regression Trees James Cussens York Centre for Complex Systems Analysis & Dept of Computer Science University of York, UK 1 Outline Problems for Lessons from Bayesian phylogeny
More informationSTA 4273H: Statistical Machine Learning
STA 4273H: Statistical Machine Learning Russ Salakhutdinov Department of Computer Science! Department of Statistical Sciences! rsalakhu@cs.toronto.edu! h0p://www.cs.utoronto.ca/~rsalakhu/ Lecture 7 Approximate
More informationLecture 6: Graphical Models: Learning
Lecture 6: Graphical Models: Learning 4F13: Machine Learning Zoubin Ghahramani and Carl Edward Rasmussen Department of Engineering, University of Cambridge February 3rd, 2010 Ghahramani & Rasmussen (CUED)
More informationOutline. 1 Introduction. 2 Problem and solution. 3 Bayesian tracking model of group debris. 4 Simulation results. 5 Conclusions
Outline 1 Introduction 2 Problem and solution 3 Bayesian tracking model of group debris 4 Simulation results 5 Conclusions Problem The limited capability of the radar can t always satisfy the detection
More informationApproximate Bayesian Computation
Approximate Bayesian Computation Michael Gutmann https://sites.google.com/site/michaelgutmann University of Helsinki and Aalto University 1st December 2015 Content Two parts: 1. The basics of approximate
More informationIntroduction to Machine Learning
Introduction to Machine Learning Brown University CSCI 1950-F, Spring 2012 Prof. Erik Sudderth Lecture 25: Markov Chain Monte Carlo (MCMC) Course Review and Advanced Topics Many figures courtesy Kevin
More informationBayesian Inference of Interactions and Associations
Bayesian Inference of Interactions and Associations Jun Liu Department of Statistics Harvard University http://www.fas.harvard.edu/~junliu Based on collaborations with Yu Zhang, Jing Zhang, Yuan Yuan,
More information[Part 2] Model Development for the Prediction of Survival Times using Longitudinal Measurements
[Part 2] Model Development for the Prediction of Survival Times using Longitudinal Measurements Aasthaa Bansal PhD Pharmaceutical Outcomes Research & Policy Program University of Washington 69 Biomarkers
More informationBeyond MCMC in fitting complex Bayesian models: The INLA method
Beyond MCMC in fitting complex Bayesian models: The INLA method Valeska Andreozzi Centre of Statistics and Applications of Lisbon University (valeska.andreozzi at fc.ul.pt) European Congress of Epidemiology
More informationMultilevel Statistical Models: 3 rd edition, 2003 Contents
Multilevel Statistical Models: 3 rd edition, 2003 Contents Preface Acknowledgements Notation Two and three level models. A general classification notation and diagram Glossary Chapter 1 An introduction
More informationLecture 2: Linear Models. Bruce Walsh lecture notes Seattle SISG -Mixed Model Course version 23 June 2011
Lecture 2: Linear Models Bruce Walsh lecture notes Seattle SISG -Mixed Model Course version 23 June 2011 1 Quick Review of the Major Points The general linear model can be written as y = X! + e y = vector
More informationBayesian Methods for Machine Learning
Bayesian Methods for Machine Learning CS 584: Big Data Analytics Material adapted from Radford Neal s tutorial (http://ftp.cs.utoronto.ca/pub/radford/bayes-tut.pdf), Zoubin Ghahramni (http://hunch.net/~coms-4771/zoubin_ghahramani_bayesian_learning.pdf),
More informationPart 2: Multivariate fmri analysis using a sparsifying spatio-temporal prior
Chalmers Machine Learning Summer School Approximate message passing and biomedicine Part 2: Multivariate fmri analysis using a sparsifying spatio-temporal prior Tom Heskes joint work with Marcel van Gerven
More informationBayesian Linear Regression
Bayesian Linear Regression Sudipto Banerjee 1 Biostatistics, School of Public Health, University of Minnesota, Minneapolis, Minnesota, U.S.A. September 15, 2010 1 Linear regression models: a Bayesian perspective
More informationDensity Estimation. Seungjin Choi
Density Estimation Seungjin Choi Department of Computer Science and Engineering Pohang University of Science and Technology 77 Cheongam-ro, Nam-gu, Pohang 37673, Korea seungjin@postech.ac.kr http://mlg.postech.ac.kr/
More informationMachine Learning Summer School
Machine Learning Summer School Lecture 3: Learning parameters and structure Zoubin Ghahramani zoubin@eng.cam.ac.uk http://learning.eng.cam.ac.uk/zoubin/ Department of Engineering University of Cambridge,
More informationCPSC 540: Machine Learning
CPSC 540: Machine Learning MCMC and Non-Parametric Bayes Mark Schmidt University of British Columbia Winter 2016 Admin I went through project proposals: Some of you got a message on Piazza. No news is
More informationNon-Parametric Bayes
Non-Parametric Bayes Mark Schmidt UBC Machine Learning Reading Group January 2016 Current Hot Topics in Machine Learning Bayesian learning includes: Gaussian processes. Approximate inference. Bayesian
More informationMCMC for big data. Geir Storvik. BigInsight lunch - May Geir Storvik MCMC for big data BigInsight lunch - May / 17
MCMC for big data Geir Storvik BigInsight lunch - May 2 2018 Geir Storvik MCMC for big data BigInsight lunch - May 2 2018 1 / 17 Outline Why ordinary MCMC is not scalable Different approaches for making
More informationProbabilistic Machine Learning
Probabilistic Machine Learning Bayesian Nets, MCMC, and more Marek Petrik 4/18/2017 Based on: P. Murphy, K. (2012). Machine Learning: A Probabilistic Perspective. Chapter 10. Conditional Independence Independent
More informationLogistic Regression. Seungjin Choi
Logistic Regression Seungjin Choi Department of Computer Science and Engineering Pohang University of Science and Technology 77 Cheongam-ro, Nam-gu, Pohang 37673, Korea seungjin@postech.ac.kr http://mlg.postech.ac.kr/
More informationEstimation in Generalized Linear Models with Heterogeneous Random Effects. Woncheol Jang Johan Lim. May 19, 2004
Estimation in Generalized Linear Models with Heterogeneous Random Effects Woncheol Jang Johan Lim May 19, 2004 Abstract The penalized quasi-likelihood (PQL) approach is the most common estimation procedure
More informationProbabilistic Graphical Models
Probabilistic Graphical Models Brown University CSCI 2950-P, Spring 2013 Prof. Erik Sudderth Lecture 13: Learning in Gaussian Graphical Models, Non-Gaussian Inference, Monte Carlo Methods Some figures
More informationLogistic regression. 11 Nov Logistic regression (EPFL) Applied Statistics 11 Nov / 20
Logistic regression 11 Nov 2010 Logistic regression (EPFL) Applied Statistics 11 Nov 2010 1 / 20 Modeling overview Want to capture important features of the relationship between a (set of) variable(s)
More informationRelated Concepts: Lecture 9 SEM, Statistical Modeling, AI, and Data Mining. I. Terminology of SEM
Lecture 9 SEM, Statistical Modeling, AI, and Data Mining I. Terminology of SEM Related Concepts: Causal Modeling Path Analysis Structural Equation Modeling Latent variables (Factors measurable, but thru
More informationan introduction to bayesian inference
with an application to network analysis http://jakehofman.com january 13, 2010 motivation would like models that: provide predictive and explanatory power are complex enough to describe observed phenomena
More informationBayesian Inference. Chapter 4: Regression and Hierarchical Models
Bayesian Inference Chapter 4: Regression and Hierarchical Models Conchi Ausín and Mike Wiper Department of Statistics Universidad Carlos III de Madrid Advanced Statistics and Data Mining Summer School
More informationCommunity Health Needs Assessment through Spatial Regression Modeling
Community Health Needs Assessment through Spatial Regression Modeling Glen D. Johnson, PhD CUNY School of Public Health glen.johnson@lehman.cuny.edu Objectives: Assess community needs with respect to particular
More informationGenerative Clustering, Topic Modeling, & Bayesian Inference
Generative Clustering, Topic Modeling, & Bayesian Inference INFO-4604, Applied Machine Learning University of Colorado Boulder December 12-14, 2017 Prof. Michael Paul Unsupervised Naïve Bayes Last week
More informationIntegrated Non-Factorized Variational Inference
Integrated Non-Factorized Variational Inference Shaobo Han, Xuejun Liao and Lawrence Carin Duke University February 27, 2014 S. Han et al. Integrated Non-Factorized Variational Inference February 27, 2014
More informationRandom Effects Models for Network Data
Random Effects Models for Network Data Peter D. Hoff 1 Working Paper no. 28 Center for Statistics and the Social Sciences University of Washington Seattle, WA 98195-4320 January 14, 2003 1 Department of
More informationSTA 4273H: Statistical Machine Learning
STA 4273H: Statistical Machine Learning Russ Salakhutdinov Department of Statistics! rsalakhu@utstat.toronto.edu! http://www.utstat.utoronto.ca/~rsalakhu/ Sidney Smith Hall, Room 6002 Lecture 3 Linear
More informationPMR Learning as Inference
Outline PMR Learning as Inference Probabilistic Modelling and Reasoning Amos Storkey Modelling 2 The Exponential Family 3 Bayesian Sets School of Informatics, University of Edinburgh Amos Storkey PMR Learning
More informationStatistical Inference and Methods
Department of Mathematics Imperial College London d.stephens@imperial.ac.uk http://stats.ma.ic.ac.uk/ das01/ 31st January 2006 Part VI Session 6: Filtering and Time to Event Data Session 6: Filtering and
More informationLogistic Regression Review Fall 2012 Recitation. September 25, 2012 TA: Selen Uguroglu
Logistic Regression Review 10-601 Fall 2012 Recitation September 25, 2012 TA: Selen Uguroglu!1 Outline Decision Theory Logistic regression Goal Loss function Inference Gradient Descent!2 Training Data
More informationPackage LBLGXE. R topics documented: July 20, Type Package
Type Package Package LBLGXE July 20, 2015 Title Bayesian Lasso for detecting Rare (or Common) Haplotype Association and their interactions with Environmental Covariates Version 1.2 Date 2015-07-09 Author
More information(5) Multi-parameter models - Gibbs sampling. ST440/540: Applied Bayesian Analysis
Summarizing a posterior Given the data and prior the posterior is determined Summarizing the posterior gives parameter estimates, intervals, and hypothesis tests Most of these computations are integrals
More informationBayesian (conditionally) conjugate inference for discrete data models. Jon Forster (University of Southampton)
Bayesian (conditionally) conjugate inference for discrete data models Jon Forster (University of Southampton) with Mark Grigsby (Procter and Gamble?) Emily Webb (Institute of Cancer Research) Table 1:
More informationThe sbgcop Package. March 9, 2007
The sbgcop Package March 9, 2007 Title Semiparametric Bayesian Gaussian copula estimation Version 0.95 Date 2007-03-09 Author Maintainer This package estimates parameters of
More informationGaussian Processes. Le Song. Machine Learning II: Advanced Topics CSE 8803ML, Spring 2012
Gaussian Processes Le Song Machine Learning II: Advanced Topics CSE 8803ML, Spring 01 Pictorial view of embedding distribution Transform the entire distribution to expected features Feature space Feature
More informationBayesian Inference and MCMC
Bayesian Inference and MCMC Aryan Arbabi Partly based on MCMC slides from CSC412 Fall 2018 1 / 18 Bayesian Inference - Motivation Consider we have a data set D = {x 1,..., x n }. E.g each x i can be the
More informationSTA 4273H: Sta-s-cal Machine Learning
STA 4273H: Sta-s-cal Machine Learning Russ Salakhutdinov Department of Computer Science! Department of Statistical Sciences! rsalakhu@cs.toronto.edu! h0p://www.cs.utoronto.ca/~rsalakhu/ Lecture 2 In our
More informationBayesian Networks: Construction, Inference, Learning and Causal Interpretation. Volker Tresp Summer 2016
Bayesian Networks: Construction, Inference, Learning and Causal Interpretation Volker Tresp Summer 2016 1 Introduction So far we were mostly concerned with supervised learning: we predicted one or several
More informationComputational Genomics. Systems biology. Putting it together: Data integration using graphical models
02-710 Computational Genomics Systems biology Putting it together: Data integration using graphical models High throughput data So far in this class we discussed several different types of high throughput
More informationSTA414/2104. Lecture 11: Gaussian Processes. Department of Statistics
STA414/2104 Lecture 11: Gaussian Processes Department of Statistics www.utstat.utoronto.ca Delivered by Mark Ebden with thanks to Russ Salakhutdinov Outline Gaussian Processes Exam review Course evaluations
More informationCoupled Hidden Markov Models: Computational Challenges
.. Coupled Hidden Markov Models: Computational Challenges Louis J. M. Aslett and Chris C. Holmes i-like Research Group University of Oxford Warwick Algorithms Seminar 7 th March 2014 ... Hidden Markov
More informationNonparmeteric Bayes & Gaussian Processes. Baback Moghaddam Machine Learning Group
Nonparmeteric Bayes & Gaussian Processes Baback Moghaddam baback@jpl.nasa.gov Machine Learning Group Outline Bayesian Inference Hierarchical Models Model Selection Parametric vs. Nonparametric Gaussian
More informationPenalized Loss functions for Bayesian Model Choice
Penalized Loss functions for Bayesian Model Choice Martyn International Agency for Research on Cancer Lyon, France 13 November 2009 The pure approach For a Bayesian purist, all uncertainty is represented
More informationBayesian Inference. Chapter 4: Regression and Hierarchical Models
Bayesian Inference Chapter 4: Regression and Hierarchical Models Conchi Ausín and Mike Wiper Department of Statistics Universidad Carlos III de Madrid Master in Business Administration and Quantitative
More informationSTA216: Generalized Linear Models. Lecture 1. Review and Introduction
STA216: Generalized Linear Models Lecture 1. Review and Introduction Let y 1,..., y n denote n independent observations on a response Treat y i as a realization of a random variable Y i In the general
More informationDefault Priors and Effcient Posterior Computation in Bayesian
Default Priors and Effcient Posterior Computation in Bayesian Factor Analysis January 16, 2010 Presented by Eric Wang, Duke University Background and Motivation A Brief Review of Parameter Expansion Literature
More informationBayesian GLMs and Metropolis-Hastings Algorithm
Bayesian GLMs and Metropolis-Hastings Algorithm We have seen that with conjugate or semi-conjugate prior distributions the Gibbs sampler can be used to sample from the posterior distribution. In situations,
More informationGeneralized Linear Models
Generalized Linear Models Advanced Methods for Data Analysis (36-402/36-608 Spring 2014 1 Generalized linear models 1.1 Introduction: two regressions So far we ve seen two canonical settings for regression.
More informationA Review of Pseudo-Marginal Markov Chain Monte Carlo
A Review of Pseudo-Marginal Markov Chain Monte Carlo Discussed by: Yizhe Zhang October 21, 2016 Outline 1 Overview 2 Paper review 3 experiment 4 conclusion Motivation & overview Notation: θ denotes the
More informationRiemann Manifold Methods in Bayesian Statistics
Ricardo Ehlers ehlers@icmc.usp.br Applied Maths and Stats University of São Paulo, Brazil Working Group in Statistical Learning University College Dublin September 2015 Bayesian inference is based on Bayes
More informationBayesian Inference of Multiple Gaussian Graphical Models
Bayesian Inference of Multiple Gaussian Graphical Models Christine Peterson,, Francesco Stingo, and Marina Vannucci February 18, 2014 Abstract In this paper, we propose a Bayesian approach to inference
More informationBayesian model selection: methodology, computation and applications
Bayesian model selection: methodology, computation and applications David Nott Department of Statistics and Applied Probability National University of Singapore Statistical Genomics Summer School Program
More informationDesign of Text Mining Experiments. Matt Taddy, University of Chicago Booth School of Business faculty.chicagobooth.edu/matt.
Design of Text Mining Experiments Matt Taddy, University of Chicago Booth School of Business faculty.chicagobooth.edu/matt.taddy/research Active Learning: a flavor of design of experiments Optimal : consider
More informationOnline appendix to On the stability of the excess sensitivity of aggregate consumption growth in the US
Online appendix to On the stability of the excess sensitivity of aggregate consumption growth in the US Gerdie Everaert 1, Lorenzo Pozzi 2, and Ruben Schoonackers 3 1 Ghent University & SHERPPA 2 Erasmus
More informationLinear Dynamical Systems
Linear Dynamical Systems Sargur N. srihari@cedar.buffalo.edu Machine Learning Course: http://www.cedar.buffalo.edu/~srihari/cse574/index.html Two Models Described by Same Graph Latent variables Observations
More informationLecture 2: From Linear Regression to Kalman Filter and Beyond
Lecture 2: From Linear Regression to Kalman Filter and Beyond Department of Biomedical Engineering and Computational Science Aalto University January 26, 2012 Contents 1 Batch and Recursive Estimation
More informationComputational statistics
Computational statistics Combinatorial optimization Thierry Denœux February 2017 Thierry Denœux Computational statistics February 2017 1 / 37 Combinatorial optimization Assume we seek the maximum of f
More informationEstimation of Parameters in Random. Effect Models with Incidence Matrix. Uncertainty
Estimation of Parameters in Random Effect Models with Incidence Matrix Uncertainty Xia Shen 1,2 and Lars Rönnegård 2,3 1 The Linnaeus Centre for Bioinformatics, Uppsala University, Uppsala, Sweden; 2 School
More informationIntroduction to Gaussian Processes
Introduction to Gaussian Processes Iain Murray murray@cs.toronto.edu CSC255, Introduction to Machine Learning, Fall 28 Dept. Computer Science, University of Toronto The problem Learn scalar function of
More informationSTA 414/2104, Spring 2014, Practice Problem Set #1
STA 44/4, Spring 4, Practice Problem Set # Note: these problems are not for credit, and not to be handed in Question : Consider a classification problem in which there are two real-valued inputs, and,
More informationOr How to select variables Using Bayesian LASSO
Or How to select variables Using Bayesian LASSO x 1 x 2 x 3 x 4 Or How to select variables Using Bayesian LASSO x 1 x 2 x 3 x 4 Or How to select variables Using Bayesian LASSO On Bayesian Variable Selection
More informationIntroduction: MLE, MAP, Bayesian reasoning (28/8/13)
STA561: Probabilistic machine learning Introduction: MLE, MAP, Bayesian reasoning (28/8/13) Lecturer: Barbara Engelhardt Scribes: K. Ulrich, J. Subramanian, N. Raval, J. O Hollaren 1 Classifiers In this
More informationSampling bias in logistic models
Sampling bias in logistic models Department of Statistics University of Chicago University of Wisconsin Oct 24, 2007 www.stat.uchicago.edu/~pmcc/reports/bias.pdf Outline Conventional regression models
More informationBayesian Inference in GLMs. Frequentists typically base inferences on MLEs, asymptotic confidence
Bayesian Inference in GLMs Frequentists typically base inferences on MLEs, asymptotic confidence limits, and log-likelihood ratio tests Bayesians base inferences on the posterior distribution of the unknowns
More informationKneib, Fahrmeir: Supplement to "Structured additive regression for categorical space-time data: A mixed model approach"
Kneib, Fahrmeir: Supplement to "Structured additive regression for categorical space-time data: A mixed model approach" Sonderforschungsbereich 386, Paper 43 (25) Online unter: http://epub.ub.uni-muenchen.de/
More information10708 Graphical Models: Homework 2
10708 Graphical Models: Homework 2 Due Monday, March 18, beginning of class Feburary 27, 2013 Instructions: There are five questions (one for extra credit) on this assignment. There is a problem involves
More informationA Brief and Friendly Introduction to Mixed-Effects Models in Linguistics
A Brief and Friendly Introduction to Mixed-Effects Models in Linguistics Cluster-specific parameters ( random effects ) Σb Parameters governing inter-cluster variability b1 b2 bm x11 x1n1 x21 x2n2 xm1
More informationDisease mapping with Gaussian processes
EUROHEIS2 Kuopio, Finland 17-18 August 2010 Aki Vehtari (former Helsinki University of Technology) Department of Biomedical Engineering and Computational Science (BECS) Acknowledgments Researchers - Jarno
More informationNonparametric Bayesian Methods (Gaussian Processes)
[70240413 Statistical Machine Learning, Spring, 2015] Nonparametric Bayesian Methods (Gaussian Processes) Jun Zhu dcszj@mail.tsinghua.edu.cn http://bigml.cs.tsinghua.edu.cn/~jun State Key Lab of Intelligent
More informationFitting Multidimensional Latent Variable Models using an Efficient Laplace Approximation
Fitting Multidimensional Latent Variable Models using an Efficient Laplace Approximation Dimitris Rizopoulos Department of Biostatistics, Erasmus University Medical Center, the Netherlands d.rizopoulos@erasmusmc.nl
More informationModeling Real Estate Data using Quantile Regression
Modeling Real Estate Data using Semiparametric Quantile Regression Department of Statistics University of Innsbruck September 9th, 2011 Overview 1 Application: 2 3 4 Hedonic regression data for house prices
More informationBayesian Image Segmentation Using MRF s Combined with Hierarchical Prior Models
Bayesian Image Segmentation Using MRF s Combined with Hierarchical Prior Models Kohta Aoki 1 and Hiroshi Nagahashi 2 1 Interdisciplinary Graduate School of Science and Engineering, Tokyo Institute of Technology
More informationarxiv:astro-ph/ v1 14 Sep 2005
For publication in Bayesian Inference and Maximum Entropy Methods, San Jose 25, K. H. Knuth, A. E. Abbas, R. D. Morris, J. P. Castle (eds.), AIP Conference Proceeding A Bayesian Analysis of Extrasolar
More informationLinear Regression. Aarti Singh. Machine Learning / Sept 27, 2010
Linear Regression Aarti Singh Machine Learning 10-701/15-781 Sept 27, 2010 Discrete to Continuous Labels Classification Sports Science News Anemic cell Healthy cell Regression X = Document Y = Topic X
More information