Probabilistic Graphical Models (Cmput 651): Hybrid Network. Matthew Brown 24/11/2008
|
|
- Patrick Gaines
- 5 years ago
- Views:
Transcription
1 Probabilistic Graphical Models (Cmput 651): Hybrid Network Matthew Brown 24/11/2008 Reading: Handout on Hybrid Networks (Ch. 13 from older version of Koller Friedman) 1
2 Space of topics Semantics Inference Learning Continuous Discrete Directed UnDirected 2
3 Outline Inference in purely continuous nets Hybrid network semantics Inference in hybrid networks 3
4 Linear Gaussian Bayesian networks (KF Definition 6.2.1) Definition: A linear Gaussian Bayesian network satisfies: all variables continuous all CPDs are linear Gaussians Example: A B C D P (A) = N (µ A, σa 2 ) P (B) = N (µ B, σb 2 ) P (C) = N (µ E C, σc 2 ) P (D A, B) = N (β D,0 + β D,1 A + β D,2 B, σd 2 ) P (E C, D) = N (β E,0 + β E,1 C + β E,2 D, σe 2 ) 4
5 Inference in linear Gaussian Bayes nets Recall: linear Gaussian Bayes nets (LGBN) equivalent to multivariate Gaussian distribution To marginalize, could convert LGBN to Gaussian marginalization trivial for Gaussian But ignores structure example LGBN: 3n 1 parameters Gaussian: n 2 +n parameters X 1 X 2... X n p(x i X i 1 ) = N (β i + α i X i 1 ; σ 2 i ) bad for large n, eg: >
6 Variable elimination Marginalize out unwanted X using integration rather than sum, as in discrete case Note: Variable elimination gives exact answers for continuous nets (not for hybrid nets) 6
7 Variable elimination example X 1 X 3 X 4 X 2 p(x 4 ) = = = P (X 1, X 2, X 3, X 4 ) X 1,X 2,X 3 P (X 1 )P (X 2 )P (X 3 X 1, X 2 )P (X 4 X 3 ) X 1,X 2,X 3 P (X 1 ) P (X 2 ) P (X 3 X 1, X 2 )P (X 4 X 3 ) X 1 X 2 X 3 Need a way to represent intermediate factors. Not Gaussian eg: conditional probabilities not (jointly) Gaussian Need elimination, product, etc. on this representation 7
8 Canonical forms (KF Handout Def n ) Definition: canonical form Also written 8
9 Canonical forms and Gaussians (KF Handout ) Canonical forms can represent Gaussians: So: 9
10 Canonical forms and Gaussians (KF Handout ) Canonical forms can represent Gaussians Other things (when K 1 not defined) eg: linear Gaussian CPDs Can also use conditional forms (multivariate linear Gaussian P(X Y) ) to represent linear Gaussian CPDs or Gaussians. 10
11 Operations on canonical forms (KF Handout ) Factor product: When scopes don t overlap, must extend them: Product of and 1st: similarly for product: 11
12 Operations on canonical forms (KF Handout ) Factor division (for belief update message passing) Note multiplying or dividing by vacuous canonical form C (0,0,0) has no effect. 12
13 Operations on canonical forms (KF Handout ) Marginalization given over set of variables {X,Y} want require K YY positive definite so that integral is finite marginal 13
14 Operations on canonical forms (KF Handout ) Conditioning given over set of variables {X,Y} want to condition on Y=y > Notice: Y no longer part of canonical form after conditioning (unlike tables). 14
15 Inference on linear Gaussian Bayesian nets (KF Handout ) Factor operations simple, closed form > Variable elimination > Sum product message passing > Belief update message passing Note on conditioning: conditioned variables disappear from canonical form unlike with factor reduction on table factors > must restrict all factors relevant to inference based on 15 evidence Y=y before doing inference
16 Inference on linear Gaussian Bayesian nets (KF Handout ) Computational performance canonical form operations polynomial in factor scope size n product & division O(n 2 ) marginalization > matrix inversion O(n 3 ) > inference in LGBNs linear in # cliques cubic in max. clique size for discrete networks factor operations on table factors exponential in scope size 16
17 Inference on linear Gaussian Bayesian nets (KF Handout ) Computational performance (cont d) for low dimensionality (small # variables), Gaussian representation can be more efficient for high dimensionality and low tree width, message passing on LGBN much more efficient 17
18 Summary Inference on linear Gaussian Bayesian nets: use canonical forms variable elimination or clique tree calibration exact efficient 18
19 Outline Inference in purely continuous nets Hybrid network semantics Inference in hybrid networks 19
20 Hybrid networks (KF 5.5.1) Hybrid networks combine discrete and continuous variables 20
21 Conditional linear Gaussian (CLG) models (KF 5.1) Definition: Given: continuous variable X with discrete parents continuous parents X has a conditional linear Gaussian CPD if for each assignment coefficients and variance such that 21
22 Conditional linear Gaussian (CLG) models (KF 5.1) Definition: A Bayesian network is a conditional linear Gaussian network if: discrete nodes have only discrete parents continuous nodes have conditional linear Gaussian CPDs continuous parents cannot have discrete children. mixture (weighted average) of Gaussians weight = probability of discrete assignment 22
23 CLG example Country Weight is CLG with continuous parent height discrete parents country and gender Weight Gender Height p(w h, c, g) = N (β c,g,0 + β c,g,1 h; σ 2 c,g) 23
24 Discrete nodes with continuous parents Option 1 hard threshold: eg: continuous X > discrete Y Y = 0 if X < 3.4 and 1 otherwise hard threshold not differentiable no gradient learning hard threshold often not realistic Option 2 soft threshold: linear sigmoid (logistic) multivariate logit NOTE: Nonlinearity! 24
25 Linear sigmoid (Logistic or soft threshold) p(y = 1 x) = exp(θt x) 1 + exp(θ T x) P(Y=1 x) x 25
26 Multivariate logit Price Trade P(trade price) Price Eg: stock trading buy (red) hold (green) sell (blue) as function of stock price l buy = 3*(price 18) l hold = 1 l sell = 3*(price 22) 26
27 Discrete node with discrete & continuous parents Continuous parents input filtered through multivariate logit Assignment to discrete parents determines coefficients for logit 27
28 Example hybrid net Price Trade Strategy stock trade (discrete) = {buy, hold, sell} parents: price (continuous), strategy (discrete) = {1 or 2} P(trade price,strategy) Price strategy 1 (reddish) l buy = 3*(price 18) l hold = 1 l sell = 3*(price 22) strategy 2 (blue/green) l buy = 3*(price 16) l hold = 1 l sell = 1*(price 26) 28
29 Outline Inference in purely continuous nets Hybrid network semantics Inference in hybrid networks Issues Non linear dependencies in continuous nets Discrete & continuous nodes: CLGs General hybrid networks 29
30 Variable elimination example (Handout Example ) Discrete D 1... D n Continuous X 1... X n ( n ) n p(d 1... D n, X 1... X n ) = p(d i ) p(x 1 D 1 ) p(x i D i, X i 1 ) i=1 i=2 p(x 2 ) = p(d 1, D 2, X 1, X 2 ) D 1,D X 1 2 = p(d 1 )p(d 2 )p(x 1 D 1 )p(x 2 D 2, X 1 ) D 1,D X 1 2 = p(d 2 ) p(x 2 D 2, X 1 ) p(x 1 D 1 )p(d 1 ) D X 1 2 D 1 > simple in principal (but see next slide) 30
31 Difficulties with inference in hybrid nets 1. must restrict representation (i.e. factors) implicit in choice to use CLGs for example 2. marginalization difficult with arbitrary hybrid nets especially with non linear dependencies among nodes continuous parent > discrete node requires non linearity! 3. intermediate factors hard to represent / work with eg: mixture of Gaussians from conditional linear Gaussian (CLG) representation > approximation necessary with hybrid nets 31
32 Difficult marginalization (KF Handout Example ) Y X P (Y ) = N (0; 1) P (X) = N (Y 2 ; 1) X non linear in Y Joint Marginal p(x, y) = 1 Z exp( y2 (x y 2 ) 2 ) p(x) = y 1 Z exp( y2 (x y 2 ) 2 ) > No analytic (closed form) solution! 32
33 Variable elimination example (Handout Example ) Discrete binary D 1... D n X 1 p(x 1 d 1 ) = N (β 1,d1 ; σ 2 1,d 1 ) X 2... X n p(x i d i, x i 1 ) = N (β i,di + α i,di x i 1 ; σ 2 i,d i ) Want P(X 2 ) P(X 1,X 2 ) is a mixture of four Gaussians, 1 / assignment to {D 1,D 2 }: Can show P(X 2 ) also a mixture of four Gaussians. not trivial to represent and work with 33
34 Discretization (KF Handout ) What about discretizing continuous variables? Usually no: typically need fine grained representation of continuous X i.e. large # bins especially where P(X) large need inference to find where P(X) large to discretize efficiently defeats the purpose > # bins usually excessively huge AND table factors suffer from curse of dimensionality exponential in Val(X) 34
35 Summary Inference in hybrid networks Difficulties with variable elimination from non linear dependencies > non Gaussian intermediate factors from mixing discrete & continuous variables > mixtures of Gaussians General approach = approximate difficult intermediate factors with Gaussians 35
36 Outline Inference in purely continuous nets Hybrid network semantics Inference in hybrid networks Issues Non linear dependencies in continuous nets Discrete & continuous nodes: CLGs General hybrid networks 36
37 Approximating intermediate factors in VE (KF Handout ) General approach: during variable elimination, when difficult intermediate factor encountered, approximate with Gaussian BUT Gaussians cannot represent: conditional distributions (CPDs) general (unnormalized) factors > must make sure to approximate only valid distributions with Gaussians eg: to eliminate X from P(X Y), must first multiply into a factor P(Y) to give p(x,y) > CPDs must be multiplied into factors in a topological ordering i.e. an ordering with parents always before children 37
38 Example (KF Handout Example ) Cliques: C 1 = {X,Y,Z}, C 2 = {Z,W} Want P(Z W=w 1 ) Variable elimination: Step 0: initialize all cliques to vacuous canonical form C(0,0,0) i.e. initial potentials not product of initial factors > C 1 s initial factors: P(X),P(Y),P(Z X,Y) 38
39 Example cont d (KF Handout Example ) Cliques: C 1 = {X,Y,Z}, C 2 = {Z,W} Want P(Z W=w 1 ) Variable elimination: Step 1: linearize P(X) i.e. approximate with Gaussian represent as canonical form then multiply into C 1 s potential (C(0,0,0) initially) Step 2: same for P(Y) could do P(Y) in step 1, then P(X) > C 1 s potential = ˆP (X, Y ) 39
40 Example cont d (KF Handout Example ) Cliques: C 1 = {X,Y,Z}, C 2 = {Z,W} Want P(Z W=w 1 ) Variable elimination: C 1 has ˆP (X, Y )P (Z X, Y ) Step 3: estimate ˆP (X, Y, Z) P (X, Y, Z) = P (X, Y )P (Z X, Y ) (represented as canonical form) ˆP (X, Y, Z) N eliminate X,Y: ˆP (Z) = ˆP (X, Y, Z) X,Y pass as message to C ˆP (Z) 2 Note: distribution 40
41 Example cont d (KF Handout Example ) Cliques: C 1 = {X,Y,Z}, C 2 = {Z,W} Want P(Z W=w 1 ) Variable elimination: C 2 has ˆP (Z)P (W Z) Step 4: estimate ˆP (W, Z) P (W, Z) = P (Z)P (W Z) (represented as canonical form) ˆP (W, Z) N Step 5: set W=w 1 pass message to C ˆP (W = w 1, Z) 1 (canonical form) Step 6: ˆP (Z W = w 1 ) = ˆP (W = w 1, Z) ˆP (Z) Note: distribution 41
42 Definition (KF Handout Def n ) Definition: A clique tree T with a root clique C r allows topological incorporation if for any variable X, the clique to which X s CPD is assigned is upstream to or equal to the cliques to which X s parents CPDs are assigned. 42
43 Approximating with Gaussians (KF Handout , ) Local approximations: Taylor series Numerical integration Global approximation 43
44 Outline Inference in purely continuous nets Hybrid network semantics Inference in hybrid networks Issues Non linear dependencies in continuous nets Discrete & continuous nodes: CLGs General hybrid networks 44
45 Inference in general hybrid nets (KF Handout ) NP hard even for polytrees mixture of exponentially many Gaussians (1 / assignment to discrete variables) eg: 2 n assignments for n binary variables even easiest case continuous nodes have at most one discrete binary parent i.e. mixture of at most two Gaussians even for easiest approximate inference on discrete binary nodes with relative error < 0.5 relative error = 0.5 is chance 45
46 Canonical tables (KF Handout Def n ) Definition: A canonical table ϕ over discrete D and continuous X has entries ϕ(d): one per assignment D=d entry ϕ(d) = canonical form C(X;K d,h d,g d ) Can represent: table factors linear Gaussians CLGs 46
47 Canonical table example discrete country, gender continuous height, weight Country Weight Gender Height Female Male Canada C(K Can,F,h Can,F,g Can,F ) C(K Can,M,h Can,M,g Can,M ) USA C(K USA,F,h USA,F,g USA,F ) C(K USA,M,h USA,M,g USA,M ) China C(K Chi,F,h Chi,F,g Chi,F ) C(K Chi,M,h Chi,M,g Chi,M ) India C(K Ind,F,h Ind,F,g Ind,F ) C(K Ind,M,h Ind,M,g Ind,M ) Germany C(K Ger,F,h Ger,F,g Ger,F ) C(K Ger,M,h Ger,M,g Ger,M ) 47
48 Operations on canonical tables (KF Handout ) Extensions of canonical form operations: Product Division Marginalization over continuous variables Marginalization over discrete variables > factor not necessarily representable with canonical table > approximate with Gaussians whenever marginalizing (in form of canonical table) (see next slide) 48
49 Marginalization example (KF Handout ) Binary D, continuous X Canonical table: Two Gaussians (blue, green) Red: sum (marginalization over D) > not Gaussian! cannot be represented by canonical table (see next slide) 49
50 Marginalization example cont d (KF Handout ) Binary D, continuous X Canonical table: Two Gaussians (blue, green) Red: Gaussian approximation to sum over blue and green 50
51 Marginalization on canonical tables (KF Handout ) Weak marginalization approximate marginal as Gaussian necessary when marginalizing across mixture of Gaussians Note: canonical tables MUST represent valid mixture Strong marginalization exact marginalize over: marginalize out continuous variables only factor over discrete only identical canonical forms 51
52 Inference in hybrid nets (KF Handout ) Cannot marginalize discrete variables > must restrict elimination order KF Handout Example A,B,C discrete; X,Y,Z continuous possible clique tree: neither leaf clique can start message passing eg: {B,X,Y} has CPDs for P(B), P(Y B,X) but not P(X) > canonical form over {X,Y} = linear Gaussian CPDs, not Gaussians > cannot marginalize out B 52
53 Strong rooted clique trees Definition: A clique C r in a clique tree is a strong root if for each clique C 1 and its upstream neighbour C 2 C 1 C 2 {continuous variables} C 1 C 2 {discrete variables} In a strongly rooted clique tree, upward pass toward strong root does not require any weak marginalization. in downward pass, all required factors present for weak marginalization to proceed Example strongly rooted clique tree (from example on previous slide): middle clique = strong root 53
54 Strong root sometimes, exist non strongly rooted clique tree that still allow inference example (refer to example 2 slide previous) Also, issue of building strongly rooted trees see KF Handout
55 Outline Inference in purely continuous nets Hybrid network semantics Inference in hybrid networks Issues Non linear dependencies in continuous nets Discrete & continuous nodes: CLGs General hybrid networks 55
56 Inference in general hybrid nets (KF Handout ) Two issues: non linear dependencies intermediate factors > marginalization on canonical tables > non canonical tabular factor solution: approximate with Gaussians (in form of canonical tables) > applies to both issues, as discussed above > allows discrete nodes with continuous parents eg: can model thermostat 56
57 Approximate methods Above, discussed variable elimination based methods Also: particle based (KF Handout 13.5) global approximate methods 57
Stat 521A Lecture 18 1
Stat 521A Lecture 18 1 Outline Cts and discrete variables (14.1) Gaussian networks (14.2) Conditional Gaussian networks (14.3) Non-linear Gaussian networks (14.4) Sampling (14.5) 2 Hybrid networks A hybrid
More informationCPSC 540: Machine Learning
CPSC 540: Machine Learning Undirected Graphical Models Mark Schmidt University of British Columbia Winter 2016 Admin Assignment 3: 2 late days to hand it in today, Thursday is final day. Assignment 4:
More informationChris Bishop s PRML Ch. 8: Graphical Models
Chris Bishop s PRML Ch. 8: Graphical Models January 24, 2008 Introduction Visualize the structure of a probabilistic model Design and motivate new models Insights into the model s properties, in particular
More informationCOS402- Artificial Intelligence Fall Lecture 10: Bayesian Networks & Exact Inference
COS402- Artificial Intelligence Fall 2015 Lecture 10: Bayesian Networks & Exact Inference Outline Logical inference and probabilistic inference Independence and conditional independence Bayes Nets Semantics
More informationCS 188: Artificial Intelligence. Bayes Nets
CS 188: Artificial Intelligence Probabilistic Inference: Enumeration, Variable Elimination, Sampling Pieter Abbeel UC Berkeley Many slides over this course adapted from Dan Klein, Stuart Russell, Andrew
More informationDynamic models 1 Kalman filters, linearization,
Koller & Friedman: Chapter 16 Jordan: Chapters 13, 15 Uri Lerner s Thesis: Chapters 3,9 Dynamic models 1 Kalman filters, linearization, Switching KFs, Assumed density filters Probabilistic Graphical Models
More informationProbabilistic Graphical Models
Probabilistic Graphical Models Lecture 12 Dynamical Models CS/CNS/EE 155 Andreas Krause Homework 3 out tonight Start early!! Announcements Project milestones due today Please email to TAs 2 Parameter learning
More informationInference in Graphical Models Variable Elimination and Message Passing Algorithm
Inference in Graphical Models Variable Elimination and Message Passing lgorithm Le Song Machine Learning II: dvanced Topics SE 8803ML, Spring 2012 onditional Independence ssumptions Local Markov ssumption
More informationBayes Nets: Independence
Bayes Nets: Independence [These slides were created by Dan Klein and Pieter Abbeel for CS188 Intro to AI at UC Berkeley. All CS188 materials are available at http://ai.berkeley.edu.] Bayes Nets A Bayes
More informationProbabilistic Graphical Models (I)
Probabilistic Graphical Models (I) Hongxin Zhang zhx@cad.zju.edu.cn State Key Lab of CAD&CG, ZJU 2015-03-31 Probabilistic Graphical Models Modeling many real-world problems => a large number of random
More informationProbabilistic Graphical Models
Probabilistic Graphical Models Lecture 11 CRFs, Exponential Family CS/CNS/EE 155 Andreas Krause Announcements Homework 2 due today Project milestones due next Monday (Nov 9) About half the work should
More informationY1 Y2 Y3 Y4 Y1 Y2 Y3 Y4 Z1 Z2 Z3 Z4
Inference: Exploiting Local Structure aphne Koller Stanford University CS228 Handout #4 We have seen that N inference exploits the network structure, in particular the conditional independence and the
More informationRepresentation. Stefano Ermon, Aditya Grover. Stanford University. Lecture 2
Representation Stefano Ermon, Aditya Grover Stanford University Lecture 2 Stefano Ermon, Aditya Grover (AI Lab) Deep Generative Models Lecture 2 1 / 32 Learning a generative model We are given a training
More informationECE521 Tutorial 11. Topic Review. ECE521 Winter Credits to Alireza Makhzani, Alex Schwing, Rich Zemel and TAs for slides. ECE521 Tutorial 11 / 4
ECE52 Tutorial Topic Review ECE52 Winter 206 Credits to Alireza Makhzani, Alex Schwing, Rich Zemel and TAs for slides ECE52 Tutorial ECE52 Winter 206 Credits to Alireza / 4 Outline K-means, PCA 2 Bayesian
More informationIntelligent Systems (AI-2)
Intelligent Systems (AI-2) Computer Science cpsc422, Lecture 11 Oct, 3, 2016 CPSC 422, Lecture 11 Slide 1 422 big picture: Where are we? Query Planning Deterministic Logics First Order Logics Ontologies
More informationAlternative Parameterizations of Markov Networks. Sargur Srihari
Alternative Parameterizations of Markov Networks Sargur srihari@cedar.buffalo.edu 1 Topics Three types of parameterization 1. Gibbs Parameterization 2. Factor Graphs 3. Log-linear Models with Energy functions
More informationCSC 412 (Lecture 4): Undirected Graphical Models
CSC 412 (Lecture 4): Undirected Graphical Models Raquel Urtasun University of Toronto Feb 2, 2016 R Urtasun (UofT) CSC 412 Feb 2, 2016 1 / 37 Today Undirected Graphical Models: Semantics of the graph:
More informationSTA 4273H: Statistical Machine Learning
STA 4273H: Statistical Machine Learning Russ Salakhutdinov Department of Statistics! rsalakhu@utstat.toronto.edu! http://www.utstat.utoronto.ca/~rsalakhu/ Sidney Smith Hall, Room 6002 Lecture 3 Linear
More informationProbabilistic Graphical Models Homework 2: Due February 24, 2014 at 4 pm
Probabilistic Graphical Models 10-708 Homework 2: Due February 24, 2014 at 4 pm Directions. This homework assignment covers the material presented in Lectures 4-8. You must complete all four problems to
More informationIntelligent Systems (AI-2)
Intelligent Systems (AI-2) Computer Science cpsc422, Lecture 19 Oct, 24, 2016 Slide Sources Raymond J. Mooney University of Texas at Austin D. Koller, Stanford CS - Probabilistic Graphical Models D. Page,
More informationBayesian Machine Learning
Bayesian Machine Learning Andrew Gordon Wilson ORIE 6741 Lecture 4 Occam s Razor, Model Construction, and Directed Graphical Models https://people.orie.cornell.edu/andrew/orie6741 Cornell University September
More informationBelief Update in CLG Bayesian Networks With Lazy Propagation
Belief Update in CLG Bayesian Networks With Lazy Propagation Anders L Madsen HUGIN Expert A/S Gasværksvej 5 9000 Aalborg, Denmark Anders.L.Madsen@hugin.com Abstract In recent years Bayesian networks (BNs)
More informationGraphical Models and Kernel Methods
Graphical Models and Kernel Methods Jerry Zhu Department of Computer Sciences University of Wisconsin Madison, USA MLSS June 17, 2014 1 / 123 Outline Graphical Models Probabilistic Inference Directed vs.
More informationOn the Relationship between Sum-Product Networks and Bayesian Networks
On the Relationship between Sum-Product Networks and Bayesian Networks International Conference on Machine Learning, 2015 Han Zhao Mazen Melibari Pascal Poupart University of Waterloo, Waterloo, ON, Canada
More informationThe Origin of Deep Learning. Lili Mou Jan, 2015
The Origin of Deep Learning Lili Mou Jan, 2015 Acknowledgment Most of the materials come from G. E. Hinton s online course. Outline Introduction Preliminary Boltzmann Machines and RBMs Deep Belief Nets
More informationProbabilistic Graphical Models
Probabilistic Graphical Models Brown University CSCI 295-P, Spring 213 Prof. Erik Sudderth Lecture 11: Inference & Learning Overview, Gaussian Graphical Models Some figures courtesy Michael Jordan s draft
More informationAn Introduction to Bayesian Machine Learning
1 An Introduction to Bayesian Machine Learning José Miguel Hernández-Lobato Department of Engineering, Cambridge University April 8, 2013 2 What is Machine Learning? The design of computational systems
More informationLocal Probabilistic Models: Continuous Variable CPDs
Local Probabilistic Models: Continuous Variable CPDs Sargur srihari@cedar.buffalo.edu 1 Topics 1. Simple discretizing loses continuity 2. Continuous Variable CPDs 3. Linear Gaussian Model Example of car
More informationLecture 9: PGM Learning
13 Oct 2014 Intro. to Stats. Machine Learning COMP SCI 4401/7401 Table of Contents I Learning parameters in MRFs 1 Learning parameters in MRFs Inference and Learning Given parameters (of potentials) and
More informationMachine Learning 4771
Machine Learning 4771 Instructor: Tony Jebara Topic 16 Undirected Graphs Undirected Separation Inferring Marginals & Conditionals Moralization Junction Trees Triangulation Undirected Graphs Separation
More informationIntelligent Systems (AI-2)
Intelligent Systems (AI-2) Computer Science cpsc422, Lecture 18 Oct, 21, 2015 Slide Sources Raymond J. Mooney University of Texas at Austin D. Koller, Stanford CS - Probabilistic Graphical Models CPSC
More informationAnnouncements. CS 188: Artificial Intelligence Fall Causality? Example: Traffic. Topology Limits Distributions. Example: Reverse Traffic
CS 188: Artificial Intelligence Fall 2008 Lecture 16: Bayes Nets III 10/23/2008 Announcements Midterms graded, up on glookup, back Tuesday W4 also graded, back in sections / box Past homeworks in return
More informationp L yi z n m x N n xi
y i z n x n N x i Overview Directed and undirected graphs Conditional independence Exact inference Latent variables and EM Variational inference Books statistical perspective Graphical Models, S. Lauritzen
More information9 Forward-backward algorithm, sum-product on factor graphs
Massachusetts Institute of Technology Department of Electrical Engineering and Computer Science 6.438 Algorithms For Inference Fall 2014 9 Forward-backward algorithm, sum-product on factor graphs The previous
More informationLocal Probability Models
Readings: K&F 3.4, 5.~5.5 Local Probability Models Lecture 3 pr 4, 2 SE 55, Statistical Methods, Spring 2 Instructor: Su-In Lee University of Washington, Seattle Outline Last time onditional parameterization
More informationIntroduction to continuous and hybrid. Bayesian networks
Introduction to continuous and hybrid Bayesian networks Joanna Ficek Supervisor: Paul Fink, M.Sc. Department of Statistics LMU January 16, 2016 Outline Introduction Gaussians Hybrid BNs Continuous children
More informationSampling Algorithms for Probabilistic Graphical models
Sampling Algorithms for Probabilistic Graphical models Vibhav Gogate University of Washington References: Chapter 12 of Probabilistic Graphical models: Principles and Techniques by Daphne Koller and Nir
More informationCS 188: Artificial Intelligence Spring Announcements
CS 188: Artificial Intelligence Spring 2011 Lecture 16: Bayes Nets IV Inference 3/28/2011 Pieter Abbeel UC Berkeley Many slides over this course adapted from Dan Klein, Stuart Russell, Andrew Moore Announcements
More informationReview: Directed Models (Bayes Nets)
X Review: Directed Models (Bayes Nets) Lecture 3: Undirected Graphical Models Sam Roweis January 2, 24 Semantics: x y z if z d-separates x and y d-separation: z d-separates x from y if along every undirected
More informationProbabilistic Graphical Models: Representation and Inference
Probabilistic Graphical Models: Representation and Inference Aaron C. Courville Université de Montréal Note: Material for the slides is taken directly from a presentation prepared by Andrew Moore 1 Overview
More informationLecture 4 October 18th
Directed and undirected graphical models Fall 2017 Lecture 4 October 18th Lecturer: Guillaume Obozinski Scribe: In this lecture, we will assume that all random variables are discrete, to keep notations
More informationPattern Recognition and Machine Learning
Christopher M. Bishop Pattern Recognition and Machine Learning ÖSpri inger Contents Preface Mathematical notation Contents vii xi xiii 1 Introduction 1 1.1 Example: Polynomial Curve Fitting 4 1.2 Probability
More informationProbabilistic Graphical Models and Bayesian Networks. Artificial Intelligence Bert Huang Virginia Tech
Probabilistic Graphical Models and Bayesian Networks Artificial Intelligence Bert Huang Virginia Tech Concept Map for Segment Probabilistic Graphical Models Probabilistic Time Series Models Particle Filters
More information17 : Markov Chain Monte Carlo
10-708: Probabilistic Graphical Models, Spring 2015 17 : Markov Chain Monte Carlo Lecturer: Eric P. Xing Scribes: Heran Lin, Bin Deng, Yun Huang 1 Review of Monte Carlo Methods 1.1 Overview Monte Carlo
More informationMachine Learning 4771
Machine Learning 4771 Instructor: Tony Jebara Topic 11 Maximum Likelihood as Bayesian Inference Maximum A Posteriori Bayesian Gaussian Estimation Why Maximum Likelihood? So far, assumed max (log) likelihood
More informationAnnouncements. Inference. Mid-term. Inference by Enumeration. Reminder: Alarm Network. Introduction to Artificial Intelligence. V22.
Introduction to Artificial Intelligence V22.0472-001 Fall 2009 Lecture 15: Bayes Nets 3 Midterms graded Assignment 2 graded Announcements Rob Fergus Dept of Computer Science, Courant Institute, NYU Slides
More informationGraphical Models and Independence Models
Graphical Models and Independence Models Yunshu Liu ASPITRG Research Group 2014-03-04 References: [1]. Steffen Lauritzen, Graphical Models, Oxford University Press, 1996 [2]. Christopher M. Bishop, Pattern
More informationStatistical Approaches to Learning and Discovery
Statistical Approaches to Learning and Discovery Graphical Models Zoubin Ghahramani & Teddy Seidenfeld zoubin@cs.cmu.edu & teddy@stat.cmu.edu CALD / CS / Statistics / Philosophy Carnegie Mellon University
More informationECE 6504: Advanced Topics in Machine Learning Probabilistic Graphical Models and Large-Scale Learning
ECE 6504: Advanced Topics in Machine Learning Probabilistic Graphical Models and Large-Scale Learning Topics Summary of Class Advanced Topics Dhruv Batra Virginia Tech HW1 Grades Mean: 28.5/38 ~= 74.9%
More informationIntelligent Systems (AI-2)
Intelligent Systems (AI-2) Computer Science cpsc422, Lecture 19 Oct, 23, 2015 Slide Sources Raymond J. Mooney University of Texas at Austin D. Koller, Stanford CS - Probabilistic Graphical Models D. Page,
More informationσ(a) = a N (x; 0, 1 2 ) dx. σ(a) = Φ(a) =
Until now we have always worked with likelihoods and prior distributions that were conjugate to each other, allowing the computation of the posterior distribution to be done in closed form. Unfortunately,
More informationSTA 4273H: Statistical Machine Learning
STA 4273H: Statistical Machine Learning Russ Salakhutdinov Department of Computer Science! Department of Statistical Sciences! rsalakhu@cs.toronto.edu! h0p://www.cs.utoronto.ca/~rsalakhu/ Lecture 7 Approximate
More informationProbabilistic Graphical Models
Probabilistic Graphical Models Lecture 9 Undirected Models CS/CNS/EE 155 Andreas Krause Announcements Homework 2 due next Wednesday (Nov 4) in class Start early!!! Project milestones due Monday (Nov 9)
More informationUndirected Graphical Models: Markov Random Fields
Undirected Graphical Models: Markov Random Fields 40-956 Advanced Topics in AI: Probabilistic Graphical Models Sharif University of Technology Soleymani Spring 2015 Markov Random Field Structure: undirected
More informationExact Inference I. Mark Peot. In this lecture we will look at issues associated with exact inference. = =
Exact Inference I Mark Peot In this lecture we will look at issues associated with exact inference 10 Queries The objective of probabilistic inference is to compute a joint distribution of a set of query
More informationUncertainty and Bayesian Networks
Uncertainty and Bayesian Networks Tutorial 3 Tutorial 3 1 Outline Uncertainty Probability Syntax and Semantics for Uncertainty Inference Independence and Bayes Rule Syntax and Semantics for Bayesian Networks
More informationBased on slides by Richard Zemel
CSC 412/2506 Winter 2018 Probabilistic Learning and Reasoning Lecture 3: Directed Graphical Models and Latent Variables Based on slides by Richard Zemel Learning outcomes What aspects of a model can we
More informationVariable Elimination (VE) Barak Sternberg
Variable Elimination (VE) Barak Sternberg Basic Ideas in VE Example 1: Let G be a Chain Bayesian Graph: X 1 X 2 X n 1 X n How would one compute P X n = k? Using the CPDs: P X 2 = x = x Val X1 P X 1 = x
More informationRecall from last time: Conditional probabilities. Lecture 2: Belief (Bayesian) networks. Bayes ball. Example (continued) Example: Inference problem
Recall from last time: Conditional probabilities Our probabilistic models will compute and manipulate conditional probabilities. Given two random variables X, Y, we denote by Lecture 2: Belief (Bayesian)
More informationChapter 16. Structured Probabilistic Models for Deep Learning
Peng et al.: Deep Learning and Practice 1 Chapter 16 Structured Probabilistic Models for Deep Learning Peng et al.: Deep Learning and Practice 2 Structured Probabilistic Models way of using graphs to describe
More informationInference in Bayesian Networks
Andrea Passerini passerini@disi.unitn.it Machine Learning Inference in graphical models Description Assume we have evidence e on the state of a subset of variables E in the model (i.e. Bayesian Network)
More informationProbabilistic Partial Evaluation: Exploiting rule structure in probabilistic inference
Probabilistic Partial Evaluation: Exploiting rule structure in probabilistic inference David Poole University of British Columbia 1 Overview Belief Networks Variable Elimination Algorithm Parent Contexts
More information4 : Exact Inference: Variable Elimination
10-708: Probabilistic Graphical Models 10-708, Spring 2014 4 : Exact Inference: Variable Elimination Lecturer: Eric P. ing Scribes: Soumya Batra, Pradeep Dasigi, Manzil Zaheer 1 Probabilistic Inference
More informationExact Inference: Clique Trees. Sargur Srihari
Exact Inference: Clique Trees Sargur srihari@cedar.buffalo.edu 1 Topics 1. Overview 2. Variable Elimination and Clique Trees 3. Message Passing: Sum-Product VE in a Clique Tree Clique-Tree Calibration
More informationProbabilistic Graphical Models
Probabilistic Graphical Models Brown University CSCI 2950-P, Spring 2013 Prof. Erik Sudderth Lecture 12: Gaussian Belief Propagation, State Space Models and Kalman Filters Guest Kalman Filter Lecture by
More informationVariable Elimination: Algorithm
Variable Elimination: Algorithm Sargur srihari@cedar.buffalo.edu 1 Topics 1. Types of Inference Algorithms 2. Variable Elimination: the Basic ideas 3. Variable Elimination Sum-Product VE Algorithm Sum-Product
More informationEE562 ARTIFICIAL INTELLIGENCE FOR ENGINEERS
EE562 ARTIFICIAL INTELLIGENCE FOR ENGINEERS Lecture 16, 6/1/2005 University of Washington, Department of Electrical Engineering Spring 2005 Instructor: Professor Jeff A. Bilmes Uncertainty & Bayesian Networks
More informationBayes Nets III: Inference
1 Hal Daumé III (me@hal3.name) Bayes Nets III: Inference Hal Daumé III Computer Science University of Maryland me@hal3.name CS 421: Introduction to Artificial Intelligence 10 Apr 2012 Many slides courtesy
More informationVariable Elimination: Algorithm
Variable Elimination: Algorithm Sargur srihari@cedar.buffalo.edu 1 Topics 1. Types of Inference Algorithms 2. Variable Elimination: the Basic ideas 3. Variable Elimination Sum-Product VE Algorithm Sum-Product
More informationAlternative Parameterizations of Markov Networks. Sargur Srihari
Alternative Parameterizations of Markov Networks Sargur srihari@cedar.buffalo.edu 1 Topics Three types of parameterization 1. Gibbs Parameterization 2. Factor Graphs 3. Log-linear Models Features (Ising,
More information6.047 / Computational Biology: Genomes, Networks, Evolution Fall 2008
MIT OpenCourseWare http://ocw.mit.edu 6.047 / 6.878 Computational Biology: Genomes, Networks, Evolution Fall 2008 For information about citing these materials or our Terms of Use, visit: http://ocw.mit.edu/terms.
More informationMachine Learning Lecture 14
Many slides adapted from B. Schiele, S. Roth, Z. Gharahmani Machine Learning Lecture 14 Undirected Graphical Models & Inference 23.06.2015 Bastian Leibe RWTH Aachen http://www.vision.rwth-aachen.de/ leibe@vision.rwth-aachen.de
More informationDEEP LEARNING CHAPTER 3 PROBABILITY & INFORMATION THEORY
DEEP LEARNING CHAPTER 3 PROBABILITY & INFORMATION THEORY OUTLINE 3.1 Why Probability? 3.2 Random Variables 3.3 Probability Distributions 3.4 Marginal Probability 3.5 Conditional Probability 3.6 The Chain
More informationGraphical Models. Lecture 3: Local Condi6onal Probability Distribu6ons. Andrew McCallum
Graphical Models Lecture 3: Local Condi6onal Probability Distribu6ons Andrew McCallum mccallum@cs.umass.edu Thanks to Noah Smith and Carlos Guestrin for some slide materials. 1 Condi6onal Probability Distribu6ons
More informationUC Berkeley Department of Electrical Engineering and Computer Science Department of Statistics. EECS 281A / STAT 241A Statistical Learning Theory
UC Berkeley Department of Electrical Engineering and Computer Science Department of Statistics EECS 281A / STAT 241A Statistical Learning Theory Solutions to Problem Set 2 Fall 2011 Issued: Wednesday,
More informationMassachusetts Institute of Technology Department of Electrical Engineering and Computer Science Algorithms For Inference Fall 2014
Massachusetts Institute of Technology Department of Electrical Engineering and Computer Science 6.438 Algorithms For Inference Fall 2014 Problem Set 3 Issued: Thursday, September 25, 2014 Due: Thursday,
More informationLecture 17: May 29, 2002
EE596 Pat. Recog. II: Introduction to Graphical Models University of Washington Spring 2000 Dept. of Electrical Engineering Lecture 17: May 29, 2002 Lecturer: Jeff ilmes Scribe: Kurt Partridge, Salvador
More informationLecture 6: Graphical Models
Lecture 6: Graphical Models Kai-Wei Chang CS @ Uniersity of Virginia kw@kwchang.net Some slides are adapted from Viek Skirmar s course on Structured Prediction 1 So far We discussed sequence labeling tasks:
More informationA graph contains a set of nodes (vertices) connected by links (edges or arcs)
BOLTZMANN MACHINES Generative Models Graphical Models A graph contains a set of nodes (vertices) connected by links (edges or arcs) In a probabilistic graphical model, each node represents a random variable,
More informationUndirected graphical models
Undirected graphical models Semantics of probabilistic models over undirected graphs Parameters of undirected models Example applications COMP-652 and ECSE-608, February 16, 2017 1 Undirected graphical
More informationInformatics 2D Reasoning and Agents Semester 2,
Informatics 2D Reasoning and Agents Semester 2, 2017 2018 Alex Lascarides alex@inf.ed.ac.uk Lecture 23 Probabilistic Reasoning with Bayesian Networks 15th March 2018 Informatics UoE Informatics 2D 1 Where
More informationLecture : Probabilistic Machine Learning
Lecture : Probabilistic Machine Learning Riashat Islam Reasoning and Learning Lab McGill University September 11, 2018 ML : Many Methods with Many Links Modelling Views of Machine Learning Machine Learning
More informationCS 2750: Machine Learning. Bayesian Networks. Prof. Adriana Kovashka University of Pittsburgh March 14, 2016
CS 2750: Machine Learning Bayesian Networks Prof. Adriana Kovashka University of Pittsburgh March 14, 2016 Plan for today and next week Today and next time: Bayesian networks (Bishop Sec. 8.1) Conditional
More informationCS281A/Stat241A Lecture 19
CS281A/Stat241A Lecture 19 p. 1/4 CS281A/Stat241A Lecture 19 Junction Tree Algorithm Peter Bartlett CS281A/Stat241A Lecture 19 p. 2/4 Announcements My office hours: Tuesday Nov 3 (today), 1-2pm, in 723
More informationGraphical Models. Lecture 10: Variable Elimina:on, con:nued. Andrew McCallum
Graphical Models Lecture 10: Variable Elimina:on, con:nued Andrew McCallum mccallum@cs.umass.edu Thanks to Noah Smith and Carlos Guestrin for some slide materials. 1 Last Time Probabilis:c inference is
More informationMarkov Networks.
Markov Networks www.biostat.wisc.edu/~dpage/cs760/ Goals for the lecture you should understand the following concepts Markov network syntax Markov network semantics Potential functions Partition function
More informationLearning Bayesian network : Given structure and completely observed data
Learning Bayesian network : Given structure and completely observed data Probabilistic Graphical Models Sharif University of Technology Spring 2017 Soleymani Learning problem Target: true distribution
More informationNonparametric Bayesian Methods (Gaussian Processes)
[70240413 Statistical Machine Learning, Spring, 2015] Nonparametric Bayesian Methods (Gaussian Processes) Jun Zhu dcszj@mail.tsinghua.edu.cn http://bigml.cs.tsinghua.edu.cn/~jun State Key Lab of Intelligent
More informationCS 343: Artificial Intelligence
CS 343: Artificial Intelligence Bayes Nets: Sampling Prof. Scott Niekum The University of Texas at Austin [These slides based on those of Dan Klein and Pieter Abbeel for CS188 Intro to AI at UC Berkeley.
More informationVariational Inference. Sargur Srihari
Variational Inference Sargur srihari@cedar.buffalo.edu 1 Plan of discussion We first describe inference with PGMs and the intractability of exact inference Then give a taxonomy of inference algorithms
More informationIntroduction to Probabilistic Graphical Models
Introduction to Probabilistic Graphical Models Sargur Srihari srihari@cedar.buffalo.edu 1 Topics 1. What are probabilistic graphical models (PGMs) 2. Use of PGMs Engineering and AI 3. Directionality in
More informationBayesian Networks. instructor: Matteo Pozzi. x 1. x 2. x 3 x 4. x 5. x 6. x 7. x 8. x 9. Lec : Urban Systems Modeling
12735: Urban Systems Modeling Lec. 09 Bayesian Networks instructor: Matteo Pozzi x 1 x 2 x 3 x 4 x 5 x 6 x 7 x 8 x 9 1 outline example of applications how to shape a problem as a BN complexity of the inference
More informationBasic Sampling Methods
Basic Sampling Methods Sargur Srihari srihari@cedar.buffalo.edu 1 1. Motivation Topics Intractability in ML How sampling can help 2. Ancestral Sampling Using BNs 3. Transforming a Uniform Distribution
More informationUndirected Graphical Models
Outline Hong Chang Institute of Computing Technology, Chinese Academy of Sciences Machine Learning Methods (Fall 2012) Outline Outline I 1 Introduction 2 Properties Properties 3 Generative vs. Conditional
More informationCS 5522: Artificial Intelligence II
CS 5522: Artificial Intelligence II Bayes Nets: Independence Instructor: Alan Ritter Ohio State University [These slides were adapted from CS188 Intro to AI at UC Berkeley. All materials available at http://ai.berkeley.edu.]
More informationIntroduction to Graphical Models
Introduction to Graphical Models The 15 th Winter School of Statistical Physics POSCO International Center & POSTECH, Pohang 2018. 1. 9 (Tue.) Yung-Kyun Noh GENERALIZATION FOR PREDICTION 2 Probabilistic
More informationInference in Hybrid Bayesian Networks with Mixtures of Truncated Exponentials
In J. Vejnarova (ed.), Proceedings of 6th Workshop on Uncertainty Processing (WUPES-2003), 47--63, VSE-Oeconomica Publishers. Inference in Hybrid Bayesian Networks with Mixtures of Truncated Exponentials
More informationConditional Independence and Factorization
Conditional Independence and Factorization Seungjin Choi Department of Computer Science and Engineering Pohang University of Science and Technology 77 Cheongam-ro, Nam-gu, Pohang 37673, Korea seungjin@postech.ac.kr
More informationWhat you need to know about Kalman Filters
Readings: K&F: 4.5, 12.2, 12.3, 12.4, 18.1, 18.2, 18.3, 18.4 Switching Kalman Filter Dynamic Bayesian Networks Graphical Models 10708 Carlos Guestrin Carnegie Mellon University November 27 th, 2006 1 What
More informationLogistic Regression Review Fall 2012 Recitation. September 25, 2012 TA: Selen Uguroglu
Logistic Regression Review 10-601 Fall 2012 Recitation September 25, 2012 TA: Selen Uguroglu!1 Outline Decision Theory Logistic regression Goal Loss function Inference Gradient Descent!2 Training Data
More information