Introduction to continuous and hybrid. Bayesian networks

Size: px
Start display at page:

Download "Introduction to continuous and hybrid. Bayesian networks"

Transcription

1 Introduction to continuous and hybrid Bayesian networks Joanna Ficek Supervisor: Paul Fink, M.Sc. Department of Statistics LMU January 16, 2016

2 Outline Introduction Gaussians Hybrid BNs Continuous children with discrete parents Discrete children with continuous parents Exponential family Entropy Relative entropy Projections I-projection M-projections Summary

3 Why? Where?

4 Bayesian networks, [access ]

5 Barbini et al. Bayesian Approach in Medicine and Health Management

6 Wenying Yan et al. Effects of Time Point Measurement on the Reconstruction of Gene Regulatory Networks, [access ]

7 Sebastiani et al., Nature Genetics 37:435,2005

8 Applications medical diagnosis gene expression data complex genetic models robot localization risk management in robotics credit scoring...

9 Continuous nodes 1. Challenges:

10 Continuous nodes 1. Challenges: no representation for all possible densities (unlike CPTs in discrete BNs) inference issues complex distributions 2. Solutions:

11 Continuous nodes 1. Challenges: no representation for all possible densities (unlike CPTs in discrete BNs) inference issues complex distributions 2. Solutions: discretization Linear Models Gaussian approximation

12 Discretization 1. Idea: continuous domain into a finite set of intervals 2. Methods: Equal Interval Width, Equal Interval Frequency,...

13 Discretization 1. Idea: continuous domain into a finite set of intervals 2. Methods: Equal Interval Width, Equal Interval Frequency,... Select y [y 1, y 2 ] x 2 P(X [x 1, x 2 ] y ) = p(x y )dy x 1

14 Discretization 1. Idea: continuous domain into a finite set of intervals 2. Methods: Equal Interval Width, Equal Interval Frequency,...

15 Discretization 1. Idea: continuous domain into a finite set of intervals 2. Methods: Equal Interval Width, Equal Interval Frequency, Limitations: loss of information trade-off: accuracy vs. computational cost O(d c )

16 Discretization 1. Idea: continuous domain into a finite set of intervals 2. Methods: Equal Interval Width, Equal Interval Frequency, Limitations: loss of information trade-off: accuracy vs. computational cost O(d c ) 4. Alternative: Linear Models

17 Linear Models Broad class of models that satisfy independence of casual influence: influence of multiple causes can be decomposed into separate influences.

18 Linear Models Broad class of models that satisfy independence of casual influence: influence of multiple causes can be decomposed into separate influences. the effect of parents (Y 1,..., Y n ) on X can be summarized via linear function f (Y 1,..., Y n ) = where w are coefficients. k w i Y i no interactions between Y i s (only through f (Y 1,..., Y n )) i=0

19 Linear Gaussian model

20 Linear Gaussian model Definition Let X be a continuous variable with continuous parents Y 1,..., Y k. We say that X has a Linear Gaussian CPD if there are β 0,..., β k and σ 2 such that: p(x y) = N (β 0 + β T y, σ 2 )

21 Linear Gaussian model Definition Let X be a continuous variable with continuous parents Y 1,..., Y k. We say that X has a Linear Gaussian CPD if there are β 0,..., β k and σ 2 such that: p(x y) = N (β 0 + β T y, σ 2 ) X = β 0 + β 1 y β k y k + ɛ

22 Linear Gaussian model IQ exams p(iq) = N (100, 15) p(e IQ) = N ( IQ, 10) = N (75, 10)

23 Linear Gaussian model IQ exams p(iq) = N (100, 15) p(e IQ) = N ( IQ, 10) = N (75, 10) Definition A Gaussian Bayesian network is a Bayesian network where all the variables are continuous and where CPDs are linear Gaussians.

24 Gaussian BN joint Gaussian Theorem Let X be a linear Gaussian of its parents Y 1,..., Y k : p(x y) = N (β 0 + β T y; σ 2 ) Assume, that Y 1,..., Y k are jointly Gaussian with distribution N (µ; Σ). Then: The distribution of X is a normal distribution p(x ) = N (µ X ; σx 2 ) where: µ X = β 0 + β T µ σ 2 X = σ2 + β T Σβ The joint distribution over {Y, X } is a normal distribution, where: k Cov[Y i, X ] = β j Σ i,j j=0

25 Gaussian BN joint Gaussian Theorem Let X be a linear Gaussian of its parents Y 1,..., Y k : p(x y) = N (β 0 + β T y; σ 2 ) Assume, that Y 1,..., Y k are jointly Gaussian with distribution N (µ; Σ). Then: The distribution of X is a normal distribution p(x ) = N (µ X ; σx 2 ) where: µ X = β 0 + β T µ σ 2 X = σ2 + β T Σβ The joint distribution over {Y, X } is a normal distribution, where: k Cov[Y i, X ] = β j Σ i,j j=0 A Gaussian Bayesian network defines a joint Gaussian distribution.

26 Gaussian BN joint Gaussian Theorem Let X = X 1,..., X n and let P be a joint Gaussian distribution over X. Given any ordering X 1,..., X n over X, we can construct a Bayesian network graph G and a Bayesian network B such that: 1. Pa G i X 1,..., X i 1 ; 2. the CPD of X i in B is a linear Gaussian of its parents; 3. G is a minimal I-map for P.

27 Probability of being accepted to a german university (MA programme) for international students form other EU-countries. accepted

28 Probability of being accepted to a german university (MA programme) for international students form other EU-countries. country of birth IQ study learned German exams accepted

29 Hybrid BNs continuous children with discrete (and continuous) parents IQ study exams

30 Hybrid BNs continuous children with discrete (and continuous) parents IQ study exams discrete children with continuous (and discrete) parents learned German exams accepted

31 Continuous children with discrete parents

32 Continuous children with discrete parents Definition Let X be a continuous variable with discrete A = A 1,..., A m and continuous Y = Y 1,..., Y k parents. We define the Conditional Linear Gaussian (CLG) model as: k p(x a, y) = N (w a,0 + w a,iy i ; σa) 2 i=1 where w are coefficients.

33 Continuous children with discrete parents Definition Let X be a continuous variable with discrete A = A 1,..., A m and continuous Y = Y 1,..., Y k parents. We define the Conditional Linear Gaussian (CLG) model as: k p(x a, y) = N (w a,0 + w a,iy i ; σa) 2 i=1 where w are coefficients. separate linear Gaussian model for each assignment to discrete parents

34 Continuous children with discrete parents Definition Let X be a continuous variable with discrete A = A 1,..., A m and continuous Y = Y 1,..., Y k parents. We define the Conditional Linear Gaussian (CLG) model as: k p(x a, y) = N (w a,0 + w a,iy i ; σa) 2 i=1 where w are coefficients. separate linear Gaussian model for each assignment to discrete parents defines a conditional Gaussian joint distribution [Lerner et al.,2001]

35 Continuous children with discrete parents IQ study exams

36 Continuous children with discrete parents IQ study exams p(iq) = N (100, 15) p(e IQ, S = s 1 ) = N ( IQ, 10) = N (85, 10)

37 Continuous children with discrete parents IQ study exams p(iq) = N (100, 15) p(e IQ, S = s 1 ) = N ( IQ, 10) = N (85, 10) p(e IQ, S = s 0 ) = N ( IQ, 12) = N (65, 12)

38 Discrete children with continuous parents

39 Discrete children with continuous parents Threshold (τ) which determines the change in discrete values

40 Discrete children with continuous parents Threshold (τ) which determines the change in discrete values X binary {x 0, x 1 } with parents Y 1,..., Y k : f (Y 1,..., Y k ) τ P(X = x 1 ) likely to be 1 f (Y 1,..., Y k ) < τ P(X = x 1 ) likely to be 0

41 Discrete children with continuous parents Threshold (τ) which determines the change in discrete values X binary {x 0, x 1 } with parents Y 1,..., Y k : f (Y 1,..., Y k ) τ P(X = x 1 ) likely to be 1 f (Y 1,..., Y k ) < τ P(X = x 1 ) likely to be 0 f (Y 1,..., Y k ) = w 0 + k i=1 w iy i

42 Discrete children with continuous parents Threshold (τ) which determines the change in discrete values X binary {x 0, x 1 } with parents Y 1,..., Y k : f (Y 1,..., Y k ) τ P(X = x 1 ) likely to be 1 f (Y 1,..., Y k ) < τ P(X = x 1 ) likely to be 0 f (Y 1,..., Y k ) = w 0 + k i=1 w iy i Definition The CPD P(X Y 1,..., Y k ) is a logistic CPD if there are k + 1 weights w 0, w 1,..., w k such that: k P(x 1 Y 1,..., Y k ) = sigmoid(w 0 + w i Y i ) i=1

43 Discrete children with continuous parents Threshold which determines the change in discrete values Augmented CLGs [Lerner et al. (2001)] Definition Let A be a discrete variable with possible values a 1,..., a m and let Y = Y 1,..., Y k denote its continuous parents. We define the CPD in augmented Conditional Linear Gaussian model as: p(a = a i y 1,..., y k ) = exp(w i,0 + k l=1 w i,ly l ) m j=1 exp(w j,0 + k s=1 w j,sy s )

44 Discrete children with continuous parents Definition A Bayesian network with all discrete variables having only discrete parents and continuous variables having a CLG CPD is a Conditional Linear Gaussian network. learned German exams accepted

45

46 Exponential family An exponential family is specified by: a sufficient statistics function τ a parameter space that is a convex set Θ of legal parameters a natural parameter function t an auxiliary measure A over X

47 Exponential family An exponential family is specified by: a sufficient statistics function τ a parameter space that is a convex set Θ of legal parameters a natural parameter function t an auxiliary measure A over X Definition An exponential family P = {P θ : θ Θ} over set of variables X, where with partition function P θ (ξ) = 1 A(ξ)exp{ t(θ), τ(ξ) } Z(θ) Z(θ) = ξ A(ξ)exp{ t(θ), τ(ξ) }

48 Univariate normal distribution τ(x) = x, x 2 (1) t(µ, σ 2 ) = µ σ 2, 1 2σ 2 (2) Z(µ, σ 2 ) = { } µ 2 2πσexp 2σ 2 (3) Then, P(x) = { 1 1 Z(µ, σ 2 exp{ t(θ), τ(x )} = exp ) 2πσ } (x µ)2 2σ 2

49 Definition Product distribution A factor in an exponential family: φ θ (ξ) = A(ξ)exp{ t(θ), τ(ξ) } Definition Let Φ 1,..., Φ k be exponential factor families. The composition of Φ 1,..., Φ k is the family Φ 1 Φ 2... Φ k parametrized by θ = θ 1 θ 2... θ k Θ 1 Θ 2... Θ k : P θ i ( ) { } φ θi (ξ) = A i (ξ) exp t(θ), τ(ξ) } i i

50 Definition Product distribution A factor in an exponential family: φ θ (ξ) = A(ξ)exp{ t(θ), τ(ξ) } Definition Let Φ 1,..., Φ k be exponential factor families. The composition of Φ 1,..., Φ k is the family Φ 1 Φ 2... Φ k parametrized by θ = θ 1 θ 2... θ k Θ 1 Θ 2... Θ k : P θ i ( ) { } φ θi (ξ) = A i (ξ) exp t(θ), τ(ξ) } i i A Bayesian network with locally normalized exponential CPDs defines an exponential family.

51 Definition Product distribution A factor in an exponential family: φ θ (ξ) = A(ξ)exp{ θ, τ(ξ) } Definition Let Φ 1,..., Φ k be exponential factor families. The composition of Φ 1,..., Φ k is the family Φ 1 Φ 2... Φ k parametrized by θ = θ 1 θ 2... θ k Θ 1 Θ 2... Θ k : P θ i ( ) { } φ θi (ξ) = A i (ξ) exp t i (θ i ), τ i (ξ) } i i A Bayesian network with locally normalized exponential CPDs defines an exponential family.

52 Entropy Measure of degree of disorder in a system (thermodynamics, R. Clausius, 1865) Shannon s Measure of Uncertainty - statistical entropy: Definition Let P(X ) be a distribution over a random variable X. The entropy of X is H P (X ) = E P [logp(x )]

53 Entropy Measure of degree of disorder in a system (thermodynamics, R. Clausius, 1865) Shannon s Measure of Uncertainty - statistical entropy: Definition Let P(X ) be a distribution over a random variable X. The entropy of X is H P (X ) = E P [logp(x )] small entropy probability mass on a few instances large entropy probability mass uniformly spread

54 Entropy Theorem Let P θ be a distribution in an exponential family defined by the functions τ and t. Then H Pθ (X ) = lnz(θ) E Pθ [τ(x )], t(θ)

55 Entropy Theorem Let P θ be a distribution in an exponential family defined by the functions τ and t. Then H Pθ (X ) = lnz(θ) E Pθ [τ(x )], t(θ) Theorem Let P(X ) = i P(X i Pa G i ) be a distribution consistent with a Bayesian network G. Then H P (X ) = i H P (X i Pa G i ) = i P(pa G i )H P (X i pa G i ) pa G i

56 Complex distribution approximation Relative entropy Measure of inaccuracy: relative entropy (Kullback-Leibler distance) Definition Let Q and P be two distributions over random variables X 1,..., X n. The relative entropy of Q and P is: where we set log(0) = 0. [ D(Q(X 1,..., X n ) P(X 1,..., X n )) = E Q log Q(X ] 1,..., X n ) P(X 1,..., X n )

57 Complex distribution approximation Relative entropy Measure of inaccuracy: relative entropy (Kullback-Leibler distance) Definition Let Q and P be two distributions over random variables X 1,..., X n. The relative entropy of Q and P is: where we set log(0) = 0. [ D(Q(X 1,..., X n ) P(X 1,..., X n )) = E Q log Q(X ] 1,..., X n ) P(X 1,..., X n ) P,Q D(Q P) 0 D(Q P) small P close to Q small loss of information P Q D(Q P) D(P Q)

58 Relative entropy Theorem Consider a distribution Q and a distribution P θ in an exponential family defined by τ and t. Then D(Q P θ ) = H Q (X ) E Q [τ(x )], t(θ) + lnz(θ)

59 Relative entropy Theorem Consider a distribution Q and a distribution P θ in an exponential family defined by τ and t. Then D(Q P θ ) = H Q (X ) E Q [τ(x )], t(θ) + lnz(θ) Theorem If P and Q are distributions over X consistent with a Bayesian network G, then D(Q P) = i Q(pa G i )D(Q(X i, pa G i )) P(X i pa G i ) pa G i

60 Projections Project a distribution P onto family of distributions Q. I-projections M-projections Q I = arg min D(Q P) Q Q Q M = arg min D(P Q) Q Q In general: Q I Q M

61 I-projections Q I = arg min D(Q P) = arg min ( H Q(X ) + E Q [ lnp(x )]) Q Q Q Q if complex graphical model high density where P is large low density where P is small penalty for low entropy some simplification of computation is possible

62 M-projections Q M = arg min D(P Q) = arg min ( H P(X ) + E P [ lnq(x )]) Q Q Q Q learning problem attempts to match the main mass of P: high density to the regions probable according to P high penalty for low density in these regions relatively high variance use of exponential form of the distribution may simplify the computation

63 M-projections Let P be a distribution over X. Theorem Let Q be an exponential family defined by τ and t. Then Q M = Q θ if there is a set of parameters θ such that E Qθ [τ(x )] = E P [τ(x )]

64 M-projections Let P be a distribution over X. Theorem Let Q be an exponential family defined by τ and t. Then Q M = Q θ if there is a set of parameters θ such that E Qθ [τ(x )] = E P [τ(x )] Theorem Let G be a Bayesian network structure. Then Q M (X ) = i P(X i Pa G i )

65 Summary 1. Continuous and hybrid BNs have many applications 2. Challenges: representation, inference 3. Solutions: discretization, Linear Models, approximation 4. Exponential family - useful form 5. True distribution unknown or complex entropy, relative entropy 6. Projections - find approximation

66 Bibliography 1. Koller D. & Friedman N., 2009, Probabilistic Graphical Models. Principles and Techniques., The MIT Press, Massachusetts, USA 2. Koller D., 2013, on-line course Probabilistic Graphical Models, 3. Lerner et al., 2001, Exact Inference in Networks with Discrete Children of Continuous Parent, p , UAI Pourret et al.,2008, Bayesian networks : a practical guide to applications, John Wiley & Sons Ltd, ISBN: , online: Bruce-Marcot-Bay.pdf

67 Bibliography 1. Friedman et al., 2000, Using Bayesian Networks to Analyze Expression Data, Journal of Computational Biology, Volume7, No. 3/4 2. Lauritzen S. & Sheehan N., 2003, Graphical Model for Genetic Analyses, Statistical Science 2003, Vol. 18, No. 4,

68 Thank you for your attention!

Ludwig-Maximilians University

Ludwig-Maximilians University Ludwig-Maximilians University Seminar Paper Introduction to continuous and hybrid Bayesian networks. Seminar: (Imprecise) Probabilistic Graphical Models Author: Joanna Ficek Supervisor: M. Sc. Paul Fink

More information

{ p if x = 1 1 p if x = 0

{ p if x = 1 1 p if x = 0 Discrete random variables Probability mass function Given a discrete random variable X taking values in X = {v 1,..., v m }, its probability mass function P : X [0, 1] is defined as: P (v i ) = Pr[X =

More information

Machine learning - HT Maximum Likelihood

Machine learning - HT Maximum Likelihood Machine learning - HT 2016 3. Maximum Likelihood Varun Kanade University of Oxford January 27, 2016 Outline Probabilistic Framework Formulate linear regression in the language of probability Introduce

More information

The Monte Carlo Method: Bayesian Networks

The Monte Carlo Method: Bayesian Networks The Method: Bayesian Networks Dieter W. Heermann Methods 2009 Dieter W. Heermann ( Methods)The Method: Bayesian Networks 2009 1 / 18 Outline 1 Bayesian Networks 2 Gene Expression Data 3 Bayesian Networks

More information

Junction Tree, BP and Variational Methods

Junction Tree, BP and Variational Methods Junction Tree, BP and Variational Methods Adrian Weller MLSALT4 Lecture Feb 21, 2018 With thanks to David Sontag (MIT) and Tony Jebara (Columbia) for use of many slides and illustrations For more information,

More information

Learning Bayesian network : Given structure and completely observed data

Learning Bayesian network : Given structure and completely observed data Learning Bayesian network : Given structure and completely observed data Probabilistic Graphical Models Sharif University of Technology Spring 2017 Soleymani Learning problem Target: true distribution

More information

Local Probabilistic Models: Continuous Variable CPDs

Local Probabilistic Models: Continuous Variable CPDs Local Probabilistic Models: Continuous Variable CPDs Sargur srihari@cedar.buffalo.edu 1 Topics 1. Simple discretizing loses continuity 2. Continuous Variable CPDs 3. Linear Gaussian Model Example of car

More information

Probabilistic Graphical Models (Cmput 651): Hybrid Network. Matthew Brown 24/11/2008

Probabilistic Graphical Models (Cmput 651): Hybrid Network. Matthew Brown 24/11/2008 Probabilistic Graphical Models (Cmput 651): Hybrid Network Matthew Brown 24/11/2008 Reading: Handout on Hybrid Networks (Ch. 13 from older version of Koller Friedman) 1 Space of topics Semantics Inference

More information

Probabilistic Graphical Models

Probabilistic Graphical Models Probabilistic Graphical Models Lecture 11 CRFs, Exponential Family CS/CNS/EE 155 Andreas Krause Announcements Homework 2 due today Project milestones due next Monday (Nov 9) About half the work should

More information

PMR Learning as Inference

PMR Learning as Inference Outline PMR Learning as Inference Probabilistic Modelling and Reasoning Amos Storkey Modelling 2 The Exponential Family 3 Bayesian Sets School of Informatics, University of Edinburgh Amos Storkey PMR Learning

More information

STA 4273H: Statistical Machine Learning

STA 4273H: Statistical Machine Learning STA 4273H: Statistical Machine Learning Russ Salakhutdinov Department of Statistics! rsalakhu@utstat.toronto.edu! http://www.utstat.utoronto.ca/~rsalakhu/ Sidney Smith Hall, Room 6002 Lecture 3 Linear

More information

Inferring Transcriptional Regulatory Networks from High-throughput Data

Inferring Transcriptional Regulatory Networks from High-throughput Data Inferring Transcriptional Regulatory Networks from High-throughput Data Lectures 9 Oct 26, 2011 CSE 527 Computational Biology, Fall 2011 Instructor: Su-In Lee TA: Christopher Miles Monday & Wednesday 12:00-1:20

More information

Variational Inference. Sargur Srihari

Variational Inference. Sargur Srihari Variational Inference Sargur srihari@cedar.buffalo.edu 1 Plan of discussion We first describe inference with PGMs and the intractability of exact inference Then give a taxonomy of inference algorithms

More information

Computational Genomics. Systems biology. Putting it together: Data integration using graphical models

Computational Genomics. Systems biology. Putting it together: Data integration using graphical models 02-710 Computational Genomics Systems biology Putting it together: Data integration using graphical models High throughput data So far in this class we discussed several different types of high throughput

More information

Graphical Models and Independence Models

Graphical Models and Independence Models Graphical Models and Independence Models Yunshu Liu ASPITRG Research Group 2014-03-04 References: [1]. Steffen Lauritzen, Graphical Models, Oxford University Press, 1996 [2]. Christopher M. Bishop, Pattern

More information

DEEP LEARNING CHAPTER 3 PROBABILITY & INFORMATION THEORY

DEEP LEARNING CHAPTER 3 PROBABILITY & INFORMATION THEORY DEEP LEARNING CHAPTER 3 PROBABILITY & INFORMATION THEORY OUTLINE 3.1 Why Probability? 3.2 Random Variables 3.3 Probability Distributions 3.4 Marginal Probability 3.5 Conditional Probability 3.6 The Chain

More information

10708 Graphical Models: Homework 2

10708 Graphical Models: Homework 2 10708 Graphical Models: Homework 2 Due Monday, March 18, beginning of class Feburary 27, 2013 Instructions: There are five questions (one for extra credit) on this assignment. There is a problem involves

More information

CSC 412 (Lecture 4): Undirected Graphical Models

CSC 412 (Lecture 4): Undirected Graphical Models CSC 412 (Lecture 4): Undirected Graphical Models Raquel Urtasun University of Toronto Feb 2, 2016 R Urtasun (UofT) CSC 412 Feb 2, 2016 1 / 37 Today Undirected Graphical Models: Semantics of the graph:

More information

Inference in Hybrid Bayesian Networks Using Mixtures of Gaussians

Inference in Hybrid Bayesian Networks Using Mixtures of Gaussians 428 SHENOY UAI 2006 Inference in Hybrid Bayesian Networks Using Mixtures of Gaussians Prakash P. Shenoy University of Kansas School of Business 1300 Sunnyside Ave, Summerfield Hall Lawrence, KS 66045-7585

More information

Directed and Undirected Graphical Models

Directed and Undirected Graphical Models Directed and Undirected Graphical Models Adrian Weller MLSALT4 Lecture Feb 26, 2016 With thanks to David Sontag (NYU) and Tony Jebara (Columbia) for use of many slides and illustrations For more information,

More information

Expectation Maximization

Expectation Maximization Expectation Maximization Seungjin Choi Department of Computer Science and Engineering Pohang University of Science and Technology 77 Cheongam-ro, Nam-gu, Pohang 37673, Korea seungjin@postech.ac.kr 1 /

More information

STA 4273H: Statistical Machine Learning

STA 4273H: Statistical Machine Learning STA 4273H: Statistical Machine Learning Russ Salakhutdinov Department of Computer Science! Department of Statistical Sciences! rsalakhu@cs.toronto.edu! h0p://www.cs.utoronto.ca/~rsalakhu/ Lecture 7 Approximate

More information

Bayesian Networks. Motivation

Bayesian Networks. Motivation Bayesian Networks Computer Sciences 760 Spring 2014 http://pages.cs.wisc.edu/~dpage/cs760/ Motivation Assume we have five Boolean variables,,,, The joint probability is,,,, How many state configurations

More information

Undirected Graphical Models

Undirected Graphical Models Outline Hong Chang Institute of Computing Technology, Chinese Academy of Sciences Machine Learning Methods (Fall 2012) Outline Outline I 1 Introduction 2 Properties Properties 3 Generative vs. Conditional

More information

CS242: Probabilistic Graphical Models Lecture 4B: Learning Tree-Structured and Directed Graphs

CS242: Probabilistic Graphical Models Lecture 4B: Learning Tree-Structured and Directed Graphs CS242: Probabilistic Graphical Models Lecture 4B: Learning Tree-Structured and Directed Graphs Professor Erik Sudderth Brown University Computer Science October 6, 2016 Some figures and materials courtesy

More information

Bayesian Dropout. Tue Herlau, Morten Morup and Mikkel N. Schmidt. Feb 20, Discussed by: Yizhe Zhang

Bayesian Dropout. Tue Herlau, Morten Morup and Mikkel N. Schmidt. Feb 20, Discussed by: Yizhe Zhang Bayesian Dropout Tue Herlau, Morten Morup and Mikkel N. Schmidt Discussed by: Yizhe Zhang Feb 20, 2016 Outline 1 Introduction 2 Model 3 Inference 4 Experiments Dropout Training stage: A unit is present

More information

An Introduction to Bayesian Machine Learning

An Introduction to Bayesian Machine Learning 1 An Introduction to Bayesian Machine Learning José Miguel Hernández-Lobato Department of Engineering, Cambridge University April 8, 2013 2 What is Machine Learning? The design of computational systems

More information

Inference in Hybrid Bayesian Networks with Mixtures of Truncated Exponentials

Inference in Hybrid Bayesian Networks with Mixtures of Truncated Exponentials In J. Vejnarova (ed.), Proceedings of 6th Workshop on Uncertainty Processing (WUPES-2003), 47--63, VSE-Oeconomica Publishers. Inference in Hybrid Bayesian Networks with Mixtures of Truncated Exponentials

More information

Expectation Propagation Algorithm

Expectation Propagation Algorithm Expectation Propagation Algorithm 1 Shuang Wang School of Electrical and Computer Engineering University of Oklahoma, Tulsa, OK, 74135 Email: {shuangwang}@ou.edu This note contains three parts. First,

More information

Nonparameteric Regression:

Nonparameteric Regression: Nonparameteric Regression: Nadaraya-Watson Kernel Regression & Gaussian Process Regression Seungjin Choi Department of Computer Science and Engineering Pohang University of Science and Technology 77 Cheongam-ro,

More information

Towards an extension of the PC algorithm to local context-specific independencies detection

Towards an extension of the PC algorithm to local context-specific independencies detection Towards an extension of the PC algorithm to local context-specific independencies detection Feb-09-2016 Outline Background: Bayesian Networks The PC algorithm Context-specific independence: from DAGs to

More information

Probabilistic Graphical Models (I)

Probabilistic Graphical Models (I) Probabilistic Graphical Models (I) Hongxin Zhang zhx@cad.zju.edu.cn State Key Lab of CAD&CG, ZJU 2015-03-31 Probabilistic Graphical Models Modeling many real-world problems => a large number of random

More information

CMPE 58K Bayesian Statistics and Machine Learning Lecture 5

CMPE 58K Bayesian Statistics and Machine Learning Lecture 5 CMPE 58K Bayesian Statistics and Machine Learning Lecture 5 Multivariate distributions: Gaussian, Bernoulli, Probability tables Department of Computer Engineering, Boğaziçi University, Istanbul, Turkey

More information

U Logo Use Guidelines

U Logo Use Guidelines Information Theory Lecture 3: Applications to Machine Learning U Logo Use Guidelines Mark Reid logo is a contemporary n of our heritage. presents our name, d and our motto: arn the nature of things. authenticity

More information

Learning in Bayesian Networks

Learning in Bayesian Networks Learning in Bayesian Networks Florian Markowetz Max-Planck-Institute for Molecular Genetics Computational Molecular Biology Berlin Berlin: 20.06.2002 1 Overview 1. Bayesian Networks Stochastic Networks

More information

Variational Message Passing. By John Winn, Christopher M. Bishop Presented by Andy Miller

Variational Message Passing. By John Winn, Christopher M. Bishop Presented by Andy Miller Variational Message Passing By John Winn, Christopher M. Bishop Presented by Andy Miller Overview Background Variational Inference Conjugate-Exponential Models Variational Message Passing Messages Univariate

More information

Large-Deviations and Applications for Learning Tree-Structured Graphical Models

Large-Deviations and Applications for Learning Tree-Structured Graphical Models Large-Deviations and Applications for Learning Tree-Structured Graphical Models Vincent Tan Stochastic Systems Group, Lab of Information and Decision Systems, Massachusetts Institute of Technology Thesis

More information

13 : Variational Inference: Loopy Belief Propagation and Mean Field

13 : Variational Inference: Loopy Belief Propagation and Mean Field 10-708: Probabilistic Graphical Models 10-708, Spring 2012 13 : Variational Inference: Loopy Belief Propagation and Mean Field Lecturer: Eric P. Xing Scribes: Peter Schulam and William Wang 1 Introduction

More information

Directed and Undirected Graphical Models

Directed and Undirected Graphical Models Directed and Undirected Davide Bacciu Dipartimento di Informatica Università di Pisa bacciu@di.unipi.it Machine Learning: Neural Networks and Advanced Models (AA2) Last Lecture Refresher Lecture Plan Directed

More information

An Introduction to Statistical and Probabilistic Linear Models

An Introduction to Statistical and Probabilistic Linear Models An Introduction to Statistical and Probabilistic Linear Models Maximilian Mozes Proseminar Data Mining Fakultät für Informatik Technische Universität München June 07, 2017 Introduction In statistical learning

More information

Naïve Bayes classification

Naïve Bayes classification Naïve Bayes classification 1 Probability theory Random variable: a variable whose possible values are numerical outcomes of a random phenomenon. Examples: A person s height, the outcome of a coin toss

More information

Lecture 6: Graphical Models: Learning

Lecture 6: Graphical Models: Learning Lecture 6: Graphical Models: Learning 4F13: Machine Learning Zoubin Ghahramani and Carl Edward Rasmussen Department of Engineering, University of Cambridge February 3rd, 2010 Ghahramani & Rasmussen (CUED)

More information

Bayesian Decision Theory Lecture 2

Bayesian Decision Theory Lecture 2 Bayesian Decision Theory Lecture 2 Jason Corso SUNY at Buffalo 14 January 2009 J. Corso (SUNY at Buffalo) Bayesian Decision Theory Lecture 2 14 January 2009 1 / 58 Overview and Plan Covering Chapter 2

More information

Bayesian Learning. HT2015: SC4 Statistical Data Mining and Machine Learning. Maximum Likelihood Principle. The Bayesian Learning Framework

Bayesian Learning. HT2015: SC4 Statistical Data Mining and Machine Learning. Maximum Likelihood Principle. The Bayesian Learning Framework HT5: SC4 Statistical Data Mining and Machine Learning Dino Sejdinovic Department of Statistics Oxford http://www.stats.ox.ac.uk/~sejdinov/sdmml.html Maximum Likelihood Principle A generative model for

More information

Lecture 9: PGM Learning

Lecture 9: PGM Learning 13 Oct 2014 Intro. to Stats. Machine Learning COMP SCI 4401/7401 Table of Contents I Learning parameters in MRFs 1 Learning parameters in MRFs Inference and Learning Given parameters (of potentials) and

More information

Being Bayesian About Network Structure:

Being Bayesian About Network Structure: Being Bayesian About Network Structure: A Bayesian Approach to Structure Discovery in Bayesian Networks Nir Friedman and Daphne Koller Machine Learning, 2003 Presented by XianXing Zhang Duke University

More information

Parametric Techniques Lecture 3

Parametric Techniques Lecture 3 Parametric Techniques Lecture 3 Jason Corso SUNY at Buffalo 22 January 2009 J. Corso (SUNY at Buffalo) Parametric Techniques Lecture 3 22 January 2009 1 / 39 Introduction In Lecture 2, we learned how to

More information

Stat 521A Lecture 18 1

Stat 521A Lecture 18 1 Stat 521A Lecture 18 1 Outline Cts and discrete variables (14.1) Gaussian networks (14.2) Conditional Gaussian networks (14.3) Non-linear Gaussian networks (14.4) Sampling (14.5) 2 Hybrid networks A hybrid

More information

Chapter 17: Undirected Graphical Models

Chapter 17: Undirected Graphical Models Chapter 17: Undirected Graphical Models The Elements of Statistical Learning Biaobin Jiang Department of Biological Sciences Purdue University bjiang@purdue.edu October 30, 2014 Biaobin Jiang (Purdue)

More information

CS Lecture 19. Exponential Families & Expectation Propagation

CS Lecture 19. Exponential Families & Expectation Propagation CS 6347 Lecture 19 Exponential Families & Expectation Propagation Discrete State Spaces We have been focusing on the case of MRFs over discrete state spaces Probability distributions over discrete spaces

More information

arxiv: v1 [cs.it] 21 Feb 2013

arxiv: v1 [cs.it] 21 Feb 2013 q-ary Compressive Sensing arxiv:30.568v [cs.it] Feb 03 Youssef Mroueh,, Lorenzo Rosasco, CBCL, CSAIL, Massachusetts Institute of Technology LCSL, Istituto Italiano di Tecnologia and IIT@MIT lab, Istituto

More information

Bayesian Learning (II)

Bayesian Learning (II) Universität Potsdam Institut für Informatik Lehrstuhl Maschinelles Lernen Bayesian Learning (II) Niels Landwehr Overview Probabilities, expected values, variance Basic concepts of Bayesian learning MAP

More information

Posterior Regularization

Posterior Regularization Posterior Regularization 1 Introduction One of the key challenges in probabilistic structured learning, is the intractability of the posterior distribution, for fast inference. There are numerous methods

More information

Representation. Stefano Ermon, Aditya Grover. Stanford University. Lecture 2

Representation. Stefano Ermon, Aditya Grover. Stanford University. Lecture 2 Representation Stefano Ermon, Aditya Grover Stanford University Lecture 2 Stefano Ermon, Aditya Grover (AI Lab) Deep Generative Models Lecture 2 1 / 32 Learning a generative model We are given a training

More information

Machine Learning Summer School

Machine Learning Summer School Machine Learning Summer School Lecture 3: Learning parameters and structure Zoubin Ghahramani zoubin@eng.cam.ac.uk http://learning.eng.cam.ac.uk/zoubin/ Department of Engineering University of Cambridge,

More information

Variational algorithms for marginal MAP

Variational algorithms for marginal MAP Variational algorithms for marginal MAP Alexander Ihler UC Irvine CIOG Workshop November 2011 Variational algorithms for marginal MAP Alexander Ihler UC Irvine CIOG Workshop November 2011 Work with Qiang

More information

Bayesian Inference Course, WTCN, UCL, March 2013

Bayesian Inference Course, WTCN, UCL, March 2013 Bayesian Course, WTCN, UCL, March 2013 Shannon (1948) asked how much information is received when we observe a specific value of the variable x? If an unlikely event occurs then one would expect the information

More information

Sum-Product Networks. STAT946 Deep Learning Guest Lecture by Pascal Poupart University of Waterloo October 17, 2017

Sum-Product Networks. STAT946 Deep Learning Guest Lecture by Pascal Poupart University of Waterloo October 17, 2017 Sum-Product Networks STAT946 Deep Learning Guest Lecture by Pascal Poupart University of Waterloo October 17, 2017 Introduction Outline What is a Sum-Product Network? Inference Applications In more depth

More information

Parametric Techniques

Parametric Techniques Parametric Techniques Jason J. Corso SUNY at Buffalo J. Corso (SUNY at Buffalo) Parametric Techniques 1 / 39 Introduction When covering Bayesian Decision Theory, we assumed the full probabilistic structure

More information

Readings: K&F: 16.3, 16.4, Graphical Models Carlos Guestrin Carnegie Mellon University October 6 th, 2008

Readings: K&F: 16.3, 16.4, Graphical Models Carlos Guestrin Carnegie Mellon University October 6 th, 2008 Readings: K&F: 16.3, 16.4, 17.3 Bayesian Param. Learning Bayesian Structure Learning Graphical Models 10708 Carlos Guestrin Carnegie Mellon University October 6 th, 2008 10-708 Carlos Guestrin 2006-2008

More information

Chapter 16. Structured Probabilistic Models for Deep Learning

Chapter 16. Structured Probabilistic Models for Deep Learning Peng et al.: Deep Learning and Practice 1 Chapter 16 Structured Probabilistic Models for Deep Learning Peng et al.: Deep Learning and Practice 2 Structured Probabilistic Models way of using graphs to describe

More information

Learning Parameters of Undirected Models. Sargur Srihari

Learning Parameters of Undirected Models. Sargur Srihari Learning Parameters of Undirected Models Sargur srihari@cedar.buffalo.edu 1 Topics Difficulties due to Global Normalization Likelihood Function Maximum Likelihood Parameter Estimation Simple and Conjugate

More information

Graphical Models and Kernel Methods

Graphical Models and Kernel Methods Graphical Models and Kernel Methods Jerry Zhu Department of Computer Sciences University of Wisconsin Madison, USA MLSS June 17, 2014 1 / 123 Outline Graphical Models Probabilistic Inference Directed vs.

More information

G8325: Variational Bayes

G8325: Variational Bayes G8325: Variational Bayes Vincent Dorie Columbia University Wednesday, November 2nd, 2011 bridge Variational University Bayes Press 2003. On-screen viewing permitted. Printing not permitted. http://www.c

More information

Naïve Bayes classification. p ij 11/15/16. Probability theory. Probability theory. Probability theory. X P (X = x i )=1 i. Marginal Probability

Naïve Bayes classification. p ij 11/15/16. Probability theory. Probability theory. Probability theory. X P (X = x i )=1 i. Marginal Probability Probability theory Naïve Bayes classification Random variable: a variable whose possible values are numerical outcomes of a random phenomenon. s: A person s height, the outcome of a coin toss Distinguish

More information

Lecture 14. Clustering, K-means, and EM

Lecture 14. Clustering, K-means, and EM Lecture 14. Clustering, K-means, and EM Prof. Alan Yuille Spring 2014 Outline 1. Clustering 2. K-means 3. EM 1 Clustering Task: Given a set of unlabeled data D = {x 1,..., x n }, we do the following: 1.

More information

Probabilistic Graphical Models

Probabilistic Graphical Models Probabilistic Graphical Models Lecture 4 Learning Bayesian Networks CS/CNS/EE 155 Andreas Krause Announcements Another TA: Hongchao Zhou Please fill out the questionnaire about recitations Homework 1 out.

More information

Probabilistic Graphical Models

Probabilistic Graphical Models Probabilistic Graphical Models Lecture 5 Bayesian Learning of Bayesian Networks CS/CNS/EE 155 Andreas Krause Announcements Recitations: Every Tuesday 4-5:30 in 243 Annenberg Homework 1 out. Due in class

More information

Bayesian Inference by Density Ratio Estimation

Bayesian Inference by Density Ratio Estimation Bayesian Inference by Density Ratio Estimation Michael Gutmann https://sites.google.com/site/michaelgutmann Institute for Adaptive and Neural Computation School of Informatics, University of Edinburgh

More information

Density Estimation. Seungjin Choi

Density Estimation. Seungjin Choi Density Estimation Seungjin Choi Department of Computer Science and Engineering Pohang University of Science and Technology 77 Cheongam-ro, Nam-gu, Pohang 37673, Korea seungjin@postech.ac.kr http://mlg.postech.ac.kr/

More information

Intelligent Systems: Reasoning and Recognition. Reasoning with Bayesian Networks

Intelligent Systems: Reasoning and Recognition. Reasoning with Bayesian Networks Intelligent Systems: Reasoning and Recognition James L. Crowley ENSIMAG 2 / MoSIG M1 Second Semester 2016/2017 Lesson 13 24 march 2017 Reasoning with Bayesian Networks Naïve Bayesian Systems...2 Example

More information

Inference in Hybrid Bayesian Networks with Mixtures of Truncated Exponentials

Inference in Hybrid Bayesian Networks with Mixtures of Truncated Exponentials Appeared in: International Journal of Approximate Reasoning, 41(3), 257--286. Inference in Hybrid Bayesian Networks with Mixtures of Truncated Exponentials Barry R. Cobb a, Prakash P. Shenoy b a Department

More information

STA 4273H: Statistical Machine Learning

STA 4273H: Statistical Machine Learning STA 4273H: Statistical Machine Learning Russ Salakhutdinov Department of Statistics! rsalakhu@utstat.toronto.edu! http://www.utstat.utoronto.ca/~rsalakhu/ Sidney Smith Hall, Room 6002 Lecture 7 Approximate

More information

Probabilistic and Bayesian Machine Learning

Probabilistic and Bayesian Machine Learning Probabilistic and Bayesian Machine Learning Day 4: Expectation and Belief Propagation Yee Whye Teh ywteh@gatsby.ucl.ac.uk Gatsby Computational Neuroscience Unit University College London http://www.gatsby.ucl.ac.uk/

More information

Bayesian Network Structure Learning and Inference Methods for Handwriting

Bayesian Network Structure Learning and Inference Methods for Handwriting Bayesian Network Structure Learning and Inference Methods for Handwriting Mukta Puri, Sargur N. Srihari and Yi Tang CEDAR, University at Buffalo, The State University of New York, Buffalo, New York, USA

More information

Machine Learning Srihari. Information Theory. Sargur N. Srihari

Machine Learning Srihari. Information Theory. Sargur N. Srihari Information Theory Sargur N. Srihari 1 Topics 1. Entropy as an Information Measure 1. Discrete variable definition Relationship to Code Length 2. Continuous Variable Differential Entropy 2. Maximum Entropy

More information

Sampling Algorithms for Probabilistic Graphical models

Sampling Algorithms for Probabilistic Graphical models Sampling Algorithms for Probabilistic Graphical models Vibhav Gogate University of Washington References: Chapter 12 of Probabilistic Graphical models: Principles and Techniques by Daphne Koller and Nir

More information

Bayesian Networks BY: MOHAMAD ALSABBAGH

Bayesian Networks BY: MOHAMAD ALSABBAGH Bayesian Networks BY: MOHAMAD ALSABBAGH Outlines Introduction Bayes Rule Bayesian Networks (BN) Representation Size of a Bayesian Network Inference via BN BN Learning Dynamic BN Introduction Conditional

More information

Probabilistic Graphical Models

Probabilistic Graphical Models Probabilistic Graphical Models David Sontag New York University Lecture 4, February 16, 2012 David Sontag (NYU) Graphical Models Lecture 4, February 16, 2012 1 / 27 Undirected graphical models Reminder

More information

ECE 6504: Advanced Topics in Machine Learning Probabilistic Graphical Models and Large-Scale Learning

ECE 6504: Advanced Topics in Machine Learning Probabilistic Graphical Models and Large-Scale Learning ECE 6504: Advanced Topics in Machine Learning Probabilistic Graphical Models and Large-Scale Learning Topics Summary of Class Advanced Topics Dhruv Batra Virginia Tech HW1 Grades Mean: 28.5/38 ~= 74.9%

More information

Using Graphs to Describe Model Structure. Sargur N. Srihari

Using Graphs to Describe Model Structure. Sargur N. Srihari Using Graphs to Describe Model Structure Sargur N. srihari@cedar.buffalo.edu 1 Topics in Structured PGMs for Deep Learning 0. Overview 1. Challenge of Unstructured Modeling 2. Using graphs to describe

More information

Bayesian Machine Learning

Bayesian Machine Learning Bayesian Machine Learning Andrew Gordon Wilson ORIE 6741 Lecture 4 Occam s Razor, Model Construction, and Directed Graphical Models https://people.orie.cornell.edu/andrew/orie6741 Cornell University September

More information

Learning Bayesian Networks (part 1) Goals for the lecture

Learning Bayesian Networks (part 1) Goals for the lecture Learning Bayesian Networks (part 1) Mark Craven and David Page Computer Scices 760 Spring 2018 www.biostat.wisc.edu/~craven/cs760/ Some ohe slides in these lectures have been adapted/borrowed from materials

More information

13 : Variational Inference: Loopy Belief Propagation

13 : Variational Inference: Loopy Belief Propagation 10-708: Probabilistic Graphical Models 10-708, Spring 2014 13 : Variational Inference: Loopy Belief Propagation Lecturer: Eric P. Xing Scribes: Rajarshi Das, Zhengzhong Liu, Dishan Gupta 1 Introduction

More information

Introduction to Machine Learning

Introduction to Machine Learning Introduction to Machine Learning Introduction to Probabilistic Methods Varun Chandola Computer Science & Engineering State University of New York at Buffalo Buffalo, NY, USA chandola@buffalo.edu Chandola@UB

More information

Iterative Markov Chain Monte Carlo Computation of Reference Priors and Minimax Risk

Iterative Markov Chain Monte Carlo Computation of Reference Priors and Minimax Risk Iterative Markov Chain Monte Carlo Computation of Reference Priors and Minimax Risk John Lafferty School of Computer Science Carnegie Mellon University Pittsburgh, PA 15213 lafferty@cs.cmu.edu Abstract

More information

T Machine Learning: Basic Principles

T Machine Learning: Basic Principles Machine Learning: Basic Principles Bayesian Networks Laboratory of Computer and Information Science (CIS) Department of Computer Science and Engineering Helsinki University of Technology (TKK) Autumn 2007

More information

ECE521 Tutorial 11. Topic Review. ECE521 Winter Credits to Alireza Makhzani, Alex Schwing, Rich Zemel and TAs for slides. ECE521 Tutorial 11 / 4

ECE521 Tutorial 11. Topic Review. ECE521 Winter Credits to Alireza Makhzani, Alex Schwing, Rich Zemel and TAs for slides. ECE521 Tutorial 11 / 4 ECE52 Tutorial Topic Review ECE52 Winter 206 Credits to Alireza Makhzani, Alex Schwing, Rich Zemel and TAs for slides ECE52 Tutorial ECE52 Winter 206 Credits to Alireza / 4 Outline K-means, PCA 2 Bayesian

More information

Probabilistic Graphical Models

Probabilistic Graphical Models Probabilistic Graphical Models Brown University CSCI 2950-P, Spring 2013 Prof. Erik Sudderth Lecture 9: Expectation Maximiation (EM) Algorithm, Learning in Undirected Graphical Models Some figures courtesy

More information

Probability. Machine Learning and Pattern Recognition. Chris Williams. School of Informatics, University of Edinburgh. August 2014

Probability. Machine Learning and Pattern Recognition. Chris Williams. School of Informatics, University of Edinburgh. August 2014 Probability Machine Learning and Pattern Recognition Chris Williams School of Informatics, University of Edinburgh August 2014 (All of the slides in this course have been adapted from previous versions

More information

Probabilistic Reasoning. (Mostly using Bayesian Networks)

Probabilistic Reasoning. (Mostly using Bayesian Networks) Probabilistic Reasoning (Mostly using Bayesian Networks) Introduction: Why probabilistic reasoning? The world is not deterministic. (Usually because information is limited.) Ways of coping with uncertainty

More information

Online Forest Density Estimation

Online Forest Density Estimation Online Forest Density Estimation Frédéric Koriche CRIL - CNRS UMR 8188, Univ. Artois koriche@cril.fr UAI 16 1 Outline 1 Probabilistic Graphical Models 2 Online Density Estimation 3 Online Forest Density

More information

CS839: Probabilistic Graphical Models. Lecture 7: Learning Fully Observed BNs. Theo Rekatsinas

CS839: Probabilistic Graphical Models. Lecture 7: Learning Fully Observed BNs. Theo Rekatsinas CS839: Probabilistic Graphical Models Lecture 7: Learning Fully Observed BNs Theo Rekatsinas 1 Exponential family: a basic building block For a numeric random variable X p(x ) =h(x)exp T T (x) A( ) = 1

More information

Bayesian Networks Introduction to Machine Learning. Matt Gormley Lecture 24 April 9, 2018

Bayesian Networks Introduction to Machine Learning. Matt Gormley Lecture 24 April 9, 2018 10-601 Introduction to Machine Learning Machine Learning Department School of Computer Science Carnegie Mellon University Bayesian Networks Matt Gormley Lecture 24 April 9, 2018 1 Homework 7: HMMs Reminders

More information

Exact model averaging with naive Bayesian classifiers

Exact model averaging with naive Bayesian classifiers Exact model averaging with naive Bayesian classifiers Denver Dash ddash@sispittedu Decision Systems Laboratory, Intelligent Systems Program, University of Pittsburgh, Pittsburgh, PA 15213 USA Gregory F

More information

Alternative Parameterizations of Markov Networks. Sargur Srihari

Alternative Parameterizations of Markov Networks. Sargur Srihari Alternative Parameterizations of Markov Networks Sargur srihari@cedar.buffalo.edu 1 Topics Three types of parameterization 1. Gibbs Parameterization 2. Factor Graphs 3. Log-linear Models with Energy functions

More information

High-dimensional graphical model selection: Practical and information-theoretic limits

High-dimensional graphical model selection: Practical and information-theoretic limits 1 High-dimensional graphical model selection: Practical and information-theoretic limits Martin Wainwright Departments of Statistics, and EECS UC Berkeley, California, USA Based on joint work with: John

More information

SOLVING STOCHASTIC PERT NETWORKS EXACTLY USING HYBRID BAYESIAN NETWORKS

SOLVING STOCHASTIC PERT NETWORKS EXACTLY USING HYBRID BAYESIAN NETWORKS SOLVING STOCHASTIC PERT NETWORKS EXACTLY USING HYBRID BAYESIAN NETWORKS Esma Nur Cinicioglu and Prakash P. Shenoy School of Business University of Kansas, Lawrence, KS 66045 USA esmanur@ku.edu, pshenoy@ku.edu

More information

Online Learning of Probabilistic Graphical Models

Online Learning of Probabilistic Graphical Models 1/34 Online Learning of Probabilistic Graphical Models Frédéric Koriche CRIL - CNRS UMR 8188, Univ. Artois koriche@cril.fr CRIL-U Nankin 2016 Probabilistic Graphical Models 2/34 Outline 1 Probabilistic

More information

Lecture 5: GPs and Streaming regression

Lecture 5: GPs and Streaming regression Lecture 5: GPs and Streaming regression Gaussian Processes Information gain Confidence intervals COMP-652 and ECSE-608, Lecture 5 - September 19, 2017 1 Recall: Non-parametric regression Input space X

More information