Probabilistic Graphical Models (Cmput 651): Hybrid Network. Matthew Brown 24/11/2008

Size: px
Start display at page:

Download "Probabilistic Graphical Models (Cmput 651): Hybrid Network. Matthew Brown 24/11/2008"

Transcription

1 Probabilistic Graphical Models (Cmput 651): Hybrid Network Matthew Brown 24/11/2008 Reading: Handout on Hybrid Networks (Ch. 13 from older version of Koller Friedman) 1

2 Space of topics Semantics Inference Learning Continuous Discrete Directed UnDirected 2

3 Outline Inference in purely continuous nets Hybrid network semantics Inference in hybrid networks 3

4 Linear Gaussian Bayesian networks (KF Definition 6.2.1) Definition: A linear Gaussian Bayesian network satisfies: all variables continuous all CPDs are linear Gaussians Example: A B C D P (A) = N (µ A, σa 2 ) P (B) = N (µ B, σb 2 ) P (C) = N (µ E C, σc 2 ) P (D A, B) = N (β D,0 + β D,1 A + β D,2 B, σd 2 ) P (E C, D) = N (β E,0 + β E,1 C + β E,2 D, σe 2 ) 4

5 Inference in linear Gaussian Bayes nets Recall: linear Gaussian Bayes nets (LGBN) equivalent to multivariate Gaussian distribution To marginalize, could convert LGBN to Gaussian marginalization trivial for Gaussian But ignores structure example LGBN: 3n 1 parameters Gaussian: n 2 +n parameters X 1 X 2... X n p(x i X i 1 ) = N (β i + α i X i 1 ; σ 2 i ) bad for large n, eg: >

6 Variable elimination Marginalize out unwanted X using integration rather than sum, as in discrete case Note: Variable elimination gives exact answers for continuous nets (not for hybrid nets) 6

7 Variable elimination example X 1 X 3 X 4 X 2 p(x 4 ) = = = P (X 1, X 2, X 3, X 4 ) X 1,X 2,X 3 P (X 1 )P (X 2 )P (X 3 X 1, X 2 )P (X 4 X 3 ) X 1,X 2,X 3 P (X 1 ) P (X 2 ) P (X 3 X 1, X 2 )P (X 4 X 3 ) X 1 X 2 X 3 Need a way to represent intermediate factors. Not Gaussian eg: conditional probabilities not (jointly) Gaussian Need elimination, product, etc. on this representation 7

8 Canonical forms (KF Handout Def n ) Definition: canonical form Also written 8

9 Canonical forms and Gaussians (KF Handout ) Canonical forms can represent Gaussians: So: 9

10 Canonical forms and Gaussians (KF Handout ) Canonical forms can represent Gaussians Other things (when K 1 not defined) eg: linear Gaussian CPDs Can also use conditional forms (multivariate linear Gaussian P(X Y) ) to represent linear Gaussian CPDs or Gaussians. 10

11 Operations on canonical forms (KF Handout ) Factor product: When scopes don t overlap, must extend them: Product of and 1st: similarly for product: 11

12 Operations on canonical forms (KF Handout ) Factor division (for belief update message passing) Note multiplying or dividing by vacuous canonical form C (0,0,0) has no effect. 12

13 Operations on canonical forms (KF Handout ) Marginalization given over set of variables {X,Y} want require K YY positive definite so that integral is finite marginal 13

14 Operations on canonical forms (KF Handout ) Conditioning given over set of variables {X,Y} want to condition on Y=y > Notice: Y no longer part of canonical form after conditioning (unlike tables). 14

15 Inference on linear Gaussian Bayesian nets (KF Handout ) Factor operations simple, closed form > Variable elimination > Sum product message passing > Belief update message passing Note on conditioning: conditioned variables disappear from canonical form unlike with factor reduction on table factors > must restrict all factors relevant to inference based on 15 evidence Y=y before doing inference

16 Inference on linear Gaussian Bayesian nets (KF Handout ) Computational performance canonical form operations polynomial in factor scope size n product & division O(n 2 ) marginalization > matrix inversion O(n 3 ) > inference in LGBNs linear in # cliques cubic in max. clique size for discrete networks factor operations on table factors exponential in scope size 16

17 Inference on linear Gaussian Bayesian nets (KF Handout ) Computational performance (cont d) for low dimensionality (small # variables), Gaussian representation can be more efficient for high dimensionality and low tree width, message passing on LGBN much more efficient 17

18 Summary Inference on linear Gaussian Bayesian nets: use canonical forms variable elimination or clique tree calibration exact efficient 18

19 Outline Inference in purely continuous nets Hybrid network semantics Inference in hybrid networks 19

20 Hybrid networks (KF 5.5.1) Hybrid networks combine discrete and continuous variables 20

21 Conditional linear Gaussian (CLG) models (KF 5.1) Definition: Given: continuous variable X with discrete parents continuous parents X has a conditional linear Gaussian CPD if for each assignment coefficients and variance such that 21

22 Conditional linear Gaussian (CLG) models (KF 5.1) Definition: A Bayesian network is a conditional linear Gaussian network if: discrete nodes have only discrete parents continuous nodes have conditional linear Gaussian CPDs continuous parents cannot have discrete children. mixture (weighted average) of Gaussians weight = probability of discrete assignment 22

23 CLG example Country Weight is CLG with continuous parent height discrete parents country and gender Weight Gender Height p(w h, c, g) = N (β c,g,0 + β c,g,1 h; σ 2 c,g) 23

24 Discrete nodes with continuous parents Option 1 hard threshold: eg: continuous X > discrete Y Y = 0 if X < 3.4 and 1 otherwise hard threshold not differentiable no gradient learning hard threshold often not realistic Option 2 soft threshold: linear sigmoid (logistic) multivariate logit NOTE: Nonlinearity! 24

25 Linear sigmoid (Logistic or soft threshold) p(y = 1 x) = exp(θt x) 1 + exp(θ T x) P(Y=1 x) x 25

26 Multivariate logit Price Trade P(trade price) Price Eg: stock trading buy (red) hold (green) sell (blue) as function of stock price l buy = 3*(price 18) l hold = 1 l sell = 3*(price 22) 26

27 Discrete node with discrete & continuous parents Continuous parents input filtered through multivariate logit Assignment to discrete parents determines coefficients for logit 27

28 Example hybrid net Price Trade Strategy stock trade (discrete) = {buy, hold, sell} parents: price (continuous), strategy (discrete) = {1 or 2} P(trade price,strategy) Price strategy 1 (reddish) l buy = 3*(price 18) l hold = 1 l sell = 3*(price 22) strategy 2 (blue/green) l buy = 3*(price 16) l hold = 1 l sell = 1*(price 26) 28

29 Outline Inference in purely continuous nets Hybrid network semantics Inference in hybrid networks Issues Non linear dependencies in continuous nets Discrete & continuous nodes: CLGs General hybrid networks 29

30 Variable elimination example (Handout Example ) Discrete D 1... D n Continuous X 1... X n ( n ) n p(d 1... D n, X 1... X n ) = p(d i ) p(x 1 D 1 ) p(x i D i, X i 1 ) i=1 i=2 p(x 2 ) = p(d 1, D 2, X 1, X 2 ) D 1,D X 1 2 = p(d 1 )p(d 2 )p(x 1 D 1 )p(x 2 D 2, X 1 ) D 1,D X 1 2 = p(d 2 ) p(x 2 D 2, X 1 ) p(x 1 D 1 )p(d 1 ) D X 1 2 D 1 > simple in principal (but see next slide) 30

31 Difficulties with inference in hybrid nets 1. must restrict representation (i.e. factors) implicit in choice to use CLGs for example 2. marginalization difficult with arbitrary hybrid nets especially with non linear dependencies among nodes continuous parent > discrete node requires non linearity! 3. intermediate factors hard to represent / work with eg: mixture of Gaussians from conditional linear Gaussian (CLG) representation > approximation necessary with hybrid nets 31

32 Difficult marginalization (KF Handout Example ) Y X P (Y ) = N (0; 1) P (X) = N (Y 2 ; 1) X non linear in Y Joint Marginal p(x, y) = 1 Z exp( y2 (x y 2 ) 2 ) p(x) = y 1 Z exp( y2 (x y 2 ) 2 ) > No analytic (closed form) solution! 32

33 Variable elimination example (Handout Example ) Discrete binary D 1... D n X 1 p(x 1 d 1 ) = N (β 1,d1 ; σ 2 1,d 1 ) X 2... X n p(x i d i, x i 1 ) = N (β i,di + α i,di x i 1 ; σ 2 i,d i ) Want P(X 2 ) P(X 1,X 2 ) is a mixture of four Gaussians, 1 / assignment to {D 1,D 2 }: Can show P(X 2 ) also a mixture of four Gaussians. not trivial to represent and work with 33

34 Discretization (KF Handout ) What about discretizing continuous variables? Usually no: typically need fine grained representation of continuous X i.e. large # bins especially where P(X) large need inference to find where P(X) large to discretize efficiently defeats the purpose > # bins usually excessively huge AND table factors suffer from curse of dimensionality exponential in Val(X) 34

35 Summary Inference in hybrid networks Difficulties with variable elimination from non linear dependencies > non Gaussian intermediate factors from mixing discrete & continuous variables > mixtures of Gaussians General approach = approximate difficult intermediate factors with Gaussians 35

36 Outline Inference in purely continuous nets Hybrid network semantics Inference in hybrid networks Issues Non linear dependencies in continuous nets Discrete & continuous nodes: CLGs General hybrid networks 36

37 Approximating intermediate factors in VE (KF Handout ) General approach: during variable elimination, when difficult intermediate factor encountered, approximate with Gaussian BUT Gaussians cannot represent: conditional distributions (CPDs) general (unnormalized) factors > must make sure to approximate only valid distributions with Gaussians eg: to eliminate X from P(X Y), must first multiply into a factor P(Y) to give p(x,y) > CPDs must be multiplied into factors in a topological ordering i.e. an ordering with parents always before children 37

38 Example (KF Handout Example ) Cliques: C 1 = {X,Y,Z}, C 2 = {Z,W} Want P(Z W=w 1 ) Variable elimination: Step 0: initialize all cliques to vacuous canonical form C(0,0,0) i.e. initial potentials not product of initial factors > C 1 s initial factors: P(X),P(Y),P(Z X,Y) 38

39 Example cont d (KF Handout Example ) Cliques: C 1 = {X,Y,Z}, C 2 = {Z,W} Want P(Z W=w 1 ) Variable elimination: Step 1: linearize P(X) i.e. approximate with Gaussian represent as canonical form then multiply into C 1 s potential (C(0,0,0) initially) Step 2: same for P(Y) could do P(Y) in step 1, then P(X) > C 1 s potential = ˆP (X, Y ) 39

40 Example cont d (KF Handout Example ) Cliques: C 1 = {X,Y,Z}, C 2 = {Z,W} Want P(Z W=w 1 ) Variable elimination: C 1 has ˆP (X, Y )P (Z X, Y ) Step 3: estimate ˆP (X, Y, Z) P (X, Y, Z) = P (X, Y )P (Z X, Y ) (represented as canonical form) ˆP (X, Y, Z) N eliminate X,Y: ˆP (Z) = ˆP (X, Y, Z) X,Y pass as message to C ˆP (Z) 2 Note: distribution 40

41 Example cont d (KF Handout Example ) Cliques: C 1 = {X,Y,Z}, C 2 = {Z,W} Want P(Z W=w 1 ) Variable elimination: C 2 has ˆP (Z)P (W Z) Step 4: estimate ˆP (W, Z) P (W, Z) = P (Z)P (W Z) (represented as canonical form) ˆP (W, Z) N Step 5: set W=w 1 pass message to C ˆP (W = w 1, Z) 1 (canonical form) Step 6: ˆP (Z W = w 1 ) = ˆP (W = w 1, Z) ˆP (Z) Note: distribution 41

42 Definition (KF Handout Def n ) Definition: A clique tree T with a root clique C r allows topological incorporation if for any variable X, the clique to which X s CPD is assigned is upstream to or equal to the cliques to which X s parents CPDs are assigned. 42

43 Approximating with Gaussians (KF Handout , ) Local approximations: Taylor series Numerical integration Global approximation 43

44 Outline Inference in purely continuous nets Hybrid network semantics Inference in hybrid networks Issues Non linear dependencies in continuous nets Discrete & continuous nodes: CLGs General hybrid networks 44

45 Inference in general hybrid nets (KF Handout ) NP hard even for polytrees mixture of exponentially many Gaussians (1 / assignment to discrete variables) eg: 2 n assignments for n binary variables even easiest case continuous nodes have at most one discrete binary parent i.e. mixture of at most two Gaussians even for easiest approximate inference on discrete binary nodes with relative error < 0.5 relative error = 0.5 is chance 45

46 Canonical tables (KF Handout Def n ) Definition: A canonical table ϕ over discrete D and continuous X has entries ϕ(d): one per assignment D=d entry ϕ(d) = canonical form C(X;K d,h d,g d ) Can represent: table factors linear Gaussians CLGs 46

47 Canonical table example discrete country, gender continuous height, weight Country Weight Gender Height Female Male Canada C(K Can,F,h Can,F,g Can,F ) C(K Can,M,h Can,M,g Can,M ) USA C(K USA,F,h USA,F,g USA,F ) C(K USA,M,h USA,M,g USA,M ) China C(K Chi,F,h Chi,F,g Chi,F ) C(K Chi,M,h Chi,M,g Chi,M ) India C(K Ind,F,h Ind,F,g Ind,F ) C(K Ind,M,h Ind,M,g Ind,M ) Germany C(K Ger,F,h Ger,F,g Ger,F ) C(K Ger,M,h Ger,M,g Ger,M ) 47

48 Operations on canonical tables (KF Handout ) Extensions of canonical form operations: Product Division Marginalization over continuous variables Marginalization over discrete variables > factor not necessarily representable with canonical table > approximate with Gaussians whenever marginalizing (in form of canonical table) (see next slide) 48

49 Marginalization example (KF Handout ) Binary D, continuous X Canonical table: Two Gaussians (blue, green) Red: sum (marginalization over D) > not Gaussian! cannot be represented by canonical table (see next slide) 49

50 Marginalization example cont d (KF Handout ) Binary D, continuous X Canonical table: Two Gaussians (blue, green) Red: Gaussian approximation to sum over blue and green 50

51 Marginalization on canonical tables (KF Handout ) Weak marginalization approximate marginal as Gaussian necessary when marginalizing across mixture of Gaussians Note: canonical tables MUST represent valid mixture Strong marginalization exact marginalize over: marginalize out continuous variables only factor over discrete only identical canonical forms 51

52 Inference in hybrid nets (KF Handout ) Cannot marginalize discrete variables > must restrict elimination order KF Handout Example A,B,C discrete; X,Y,Z continuous possible clique tree: neither leaf clique can start message passing eg: {B,X,Y} has CPDs for P(B), P(Y B,X) but not P(X) > canonical form over {X,Y} = linear Gaussian CPDs, not Gaussians > cannot marginalize out B 52

53 Strong rooted clique trees Definition: A clique C r in a clique tree is a strong root if for each clique C 1 and its upstream neighbour C 2 C 1 C 2 {continuous variables} C 1 C 2 {discrete variables} In a strongly rooted clique tree, upward pass toward strong root does not require any weak marginalization. in downward pass, all required factors present for weak marginalization to proceed Example strongly rooted clique tree (from example on previous slide): middle clique = strong root 53

54 Strong root sometimes, exist non strongly rooted clique tree that still allow inference example (refer to example 2 slide previous) Also, issue of building strongly rooted trees see KF Handout

55 Outline Inference in purely continuous nets Hybrid network semantics Inference in hybrid networks Issues Non linear dependencies in continuous nets Discrete & continuous nodes: CLGs General hybrid networks 55

56 Inference in general hybrid nets (KF Handout ) Two issues: non linear dependencies intermediate factors > marginalization on canonical tables > non canonical tabular factor solution: approximate with Gaussians (in form of canonical tables) > applies to both issues, as discussed above > allows discrete nodes with continuous parents eg: can model thermostat 56

57 Approximate methods Above, discussed variable elimination based methods Also: particle based (KF Handout 13.5) global approximate methods 57

Stat 521A Lecture 18 1

Stat 521A Lecture 18 1 Stat 521A Lecture 18 1 Outline Cts and discrete variables (14.1) Gaussian networks (14.2) Conditional Gaussian networks (14.3) Non-linear Gaussian networks (14.4) Sampling (14.5) 2 Hybrid networks A hybrid

More information

CPSC 540: Machine Learning

CPSC 540: Machine Learning CPSC 540: Machine Learning Undirected Graphical Models Mark Schmidt University of British Columbia Winter 2016 Admin Assignment 3: 2 late days to hand it in today, Thursday is final day. Assignment 4:

More information

Chris Bishop s PRML Ch. 8: Graphical Models

Chris Bishop s PRML Ch. 8: Graphical Models Chris Bishop s PRML Ch. 8: Graphical Models January 24, 2008 Introduction Visualize the structure of a probabilistic model Design and motivate new models Insights into the model s properties, in particular

More information

COS402- Artificial Intelligence Fall Lecture 10: Bayesian Networks & Exact Inference

COS402- Artificial Intelligence Fall Lecture 10: Bayesian Networks & Exact Inference COS402- Artificial Intelligence Fall 2015 Lecture 10: Bayesian Networks & Exact Inference Outline Logical inference and probabilistic inference Independence and conditional independence Bayes Nets Semantics

More information

CS 188: Artificial Intelligence. Bayes Nets

CS 188: Artificial Intelligence. Bayes Nets CS 188: Artificial Intelligence Probabilistic Inference: Enumeration, Variable Elimination, Sampling Pieter Abbeel UC Berkeley Many slides over this course adapted from Dan Klein, Stuart Russell, Andrew

More information

Dynamic models 1 Kalman filters, linearization,

Dynamic models 1 Kalman filters, linearization, Koller & Friedman: Chapter 16 Jordan: Chapters 13, 15 Uri Lerner s Thesis: Chapters 3,9 Dynamic models 1 Kalman filters, linearization, Switching KFs, Assumed density filters Probabilistic Graphical Models

More information

Probabilistic Graphical Models

Probabilistic Graphical Models Probabilistic Graphical Models Lecture 12 Dynamical Models CS/CNS/EE 155 Andreas Krause Homework 3 out tonight Start early!! Announcements Project milestones due today Please email to TAs 2 Parameter learning

More information

Inference in Graphical Models Variable Elimination and Message Passing Algorithm

Inference in Graphical Models Variable Elimination and Message Passing Algorithm Inference in Graphical Models Variable Elimination and Message Passing lgorithm Le Song Machine Learning II: dvanced Topics SE 8803ML, Spring 2012 onditional Independence ssumptions Local Markov ssumption

More information

Bayes Nets: Independence

Bayes Nets: Independence Bayes Nets: Independence [These slides were created by Dan Klein and Pieter Abbeel for CS188 Intro to AI at UC Berkeley. All CS188 materials are available at http://ai.berkeley.edu.] Bayes Nets A Bayes

More information

Probabilistic Graphical Models (I)

Probabilistic Graphical Models (I) Probabilistic Graphical Models (I) Hongxin Zhang zhx@cad.zju.edu.cn State Key Lab of CAD&CG, ZJU 2015-03-31 Probabilistic Graphical Models Modeling many real-world problems => a large number of random

More information

Probabilistic Graphical Models

Probabilistic Graphical Models Probabilistic Graphical Models Lecture 11 CRFs, Exponential Family CS/CNS/EE 155 Andreas Krause Announcements Homework 2 due today Project milestones due next Monday (Nov 9) About half the work should

More information

Y1 Y2 Y3 Y4 Y1 Y2 Y3 Y4 Z1 Z2 Z3 Z4

Y1 Y2 Y3 Y4 Y1 Y2 Y3 Y4 Z1 Z2 Z3 Z4 Inference: Exploiting Local Structure aphne Koller Stanford University CS228 Handout #4 We have seen that N inference exploits the network structure, in particular the conditional independence and the

More information

Representation. Stefano Ermon, Aditya Grover. Stanford University. Lecture 2

Representation. Stefano Ermon, Aditya Grover. Stanford University. Lecture 2 Representation Stefano Ermon, Aditya Grover Stanford University Lecture 2 Stefano Ermon, Aditya Grover (AI Lab) Deep Generative Models Lecture 2 1 / 32 Learning a generative model We are given a training

More information

ECE521 Tutorial 11. Topic Review. ECE521 Winter Credits to Alireza Makhzani, Alex Schwing, Rich Zemel and TAs for slides. ECE521 Tutorial 11 / 4

ECE521 Tutorial 11. Topic Review. ECE521 Winter Credits to Alireza Makhzani, Alex Schwing, Rich Zemel and TAs for slides. ECE521 Tutorial 11 / 4 ECE52 Tutorial Topic Review ECE52 Winter 206 Credits to Alireza Makhzani, Alex Schwing, Rich Zemel and TAs for slides ECE52 Tutorial ECE52 Winter 206 Credits to Alireza / 4 Outline K-means, PCA 2 Bayesian

More information

Intelligent Systems (AI-2)

Intelligent Systems (AI-2) Intelligent Systems (AI-2) Computer Science cpsc422, Lecture 11 Oct, 3, 2016 CPSC 422, Lecture 11 Slide 1 422 big picture: Where are we? Query Planning Deterministic Logics First Order Logics Ontologies

More information

Alternative Parameterizations of Markov Networks. Sargur Srihari

Alternative Parameterizations of Markov Networks. Sargur Srihari Alternative Parameterizations of Markov Networks Sargur srihari@cedar.buffalo.edu 1 Topics Three types of parameterization 1. Gibbs Parameterization 2. Factor Graphs 3. Log-linear Models with Energy functions

More information

CSC 412 (Lecture 4): Undirected Graphical Models

CSC 412 (Lecture 4): Undirected Graphical Models CSC 412 (Lecture 4): Undirected Graphical Models Raquel Urtasun University of Toronto Feb 2, 2016 R Urtasun (UofT) CSC 412 Feb 2, 2016 1 / 37 Today Undirected Graphical Models: Semantics of the graph:

More information

STA 4273H: Statistical Machine Learning

STA 4273H: Statistical Machine Learning STA 4273H: Statistical Machine Learning Russ Salakhutdinov Department of Statistics! rsalakhu@utstat.toronto.edu! http://www.utstat.utoronto.ca/~rsalakhu/ Sidney Smith Hall, Room 6002 Lecture 3 Linear

More information

Probabilistic Graphical Models Homework 2: Due February 24, 2014 at 4 pm

Probabilistic Graphical Models Homework 2: Due February 24, 2014 at 4 pm Probabilistic Graphical Models 10-708 Homework 2: Due February 24, 2014 at 4 pm Directions. This homework assignment covers the material presented in Lectures 4-8. You must complete all four problems to

More information

Intelligent Systems (AI-2)

Intelligent Systems (AI-2) Intelligent Systems (AI-2) Computer Science cpsc422, Lecture 19 Oct, 24, 2016 Slide Sources Raymond J. Mooney University of Texas at Austin D. Koller, Stanford CS - Probabilistic Graphical Models D. Page,

More information

Bayesian Machine Learning

Bayesian Machine Learning Bayesian Machine Learning Andrew Gordon Wilson ORIE 6741 Lecture 4 Occam s Razor, Model Construction, and Directed Graphical Models https://people.orie.cornell.edu/andrew/orie6741 Cornell University September

More information

Belief Update in CLG Bayesian Networks With Lazy Propagation

Belief Update in CLG Bayesian Networks With Lazy Propagation Belief Update in CLG Bayesian Networks With Lazy Propagation Anders L Madsen HUGIN Expert A/S Gasværksvej 5 9000 Aalborg, Denmark Anders.L.Madsen@hugin.com Abstract In recent years Bayesian networks (BNs)

More information

Graphical Models and Kernel Methods

Graphical Models and Kernel Methods Graphical Models and Kernel Methods Jerry Zhu Department of Computer Sciences University of Wisconsin Madison, USA MLSS June 17, 2014 1 / 123 Outline Graphical Models Probabilistic Inference Directed vs.

More information

On the Relationship between Sum-Product Networks and Bayesian Networks

On the Relationship between Sum-Product Networks and Bayesian Networks On the Relationship between Sum-Product Networks and Bayesian Networks International Conference on Machine Learning, 2015 Han Zhao Mazen Melibari Pascal Poupart University of Waterloo, Waterloo, ON, Canada

More information

The Origin of Deep Learning. Lili Mou Jan, 2015

The Origin of Deep Learning. Lili Mou Jan, 2015 The Origin of Deep Learning Lili Mou Jan, 2015 Acknowledgment Most of the materials come from G. E. Hinton s online course. Outline Introduction Preliminary Boltzmann Machines and RBMs Deep Belief Nets

More information

Probabilistic Graphical Models

Probabilistic Graphical Models Probabilistic Graphical Models Brown University CSCI 295-P, Spring 213 Prof. Erik Sudderth Lecture 11: Inference & Learning Overview, Gaussian Graphical Models Some figures courtesy Michael Jordan s draft

More information

An Introduction to Bayesian Machine Learning

An Introduction to Bayesian Machine Learning 1 An Introduction to Bayesian Machine Learning José Miguel Hernández-Lobato Department of Engineering, Cambridge University April 8, 2013 2 What is Machine Learning? The design of computational systems

More information

Local Probabilistic Models: Continuous Variable CPDs

Local Probabilistic Models: Continuous Variable CPDs Local Probabilistic Models: Continuous Variable CPDs Sargur srihari@cedar.buffalo.edu 1 Topics 1. Simple discretizing loses continuity 2. Continuous Variable CPDs 3. Linear Gaussian Model Example of car

More information

Lecture 9: PGM Learning

Lecture 9: PGM Learning 13 Oct 2014 Intro. to Stats. Machine Learning COMP SCI 4401/7401 Table of Contents I Learning parameters in MRFs 1 Learning parameters in MRFs Inference and Learning Given parameters (of potentials) and

More information

Machine Learning 4771

Machine Learning 4771 Machine Learning 4771 Instructor: Tony Jebara Topic 16 Undirected Graphs Undirected Separation Inferring Marginals & Conditionals Moralization Junction Trees Triangulation Undirected Graphs Separation

More information

Intelligent Systems (AI-2)

Intelligent Systems (AI-2) Intelligent Systems (AI-2) Computer Science cpsc422, Lecture 18 Oct, 21, 2015 Slide Sources Raymond J. Mooney University of Texas at Austin D. Koller, Stanford CS - Probabilistic Graphical Models CPSC

More information

Announcements. CS 188: Artificial Intelligence Fall Causality? Example: Traffic. Topology Limits Distributions. Example: Reverse Traffic

Announcements. CS 188: Artificial Intelligence Fall Causality? Example: Traffic. Topology Limits Distributions. Example: Reverse Traffic CS 188: Artificial Intelligence Fall 2008 Lecture 16: Bayes Nets III 10/23/2008 Announcements Midterms graded, up on glookup, back Tuesday W4 also graded, back in sections / box Past homeworks in return

More information

p L yi z n m x N n xi

p L yi z n m x N n xi y i z n x n N x i Overview Directed and undirected graphs Conditional independence Exact inference Latent variables and EM Variational inference Books statistical perspective Graphical Models, S. Lauritzen

More information

9 Forward-backward algorithm, sum-product on factor graphs

9 Forward-backward algorithm, sum-product on factor graphs Massachusetts Institute of Technology Department of Electrical Engineering and Computer Science 6.438 Algorithms For Inference Fall 2014 9 Forward-backward algorithm, sum-product on factor graphs The previous

More information

Local Probability Models

Local Probability Models Readings: K&F 3.4, 5.~5.5 Local Probability Models Lecture 3 pr 4, 2 SE 55, Statistical Methods, Spring 2 Instructor: Su-In Lee University of Washington, Seattle Outline Last time onditional parameterization

More information

Introduction to continuous and hybrid. Bayesian networks

Introduction to continuous and hybrid. Bayesian networks Introduction to continuous and hybrid Bayesian networks Joanna Ficek Supervisor: Paul Fink, M.Sc. Department of Statistics LMU January 16, 2016 Outline Introduction Gaussians Hybrid BNs Continuous children

More information

Sampling Algorithms for Probabilistic Graphical models

Sampling Algorithms for Probabilistic Graphical models Sampling Algorithms for Probabilistic Graphical models Vibhav Gogate University of Washington References: Chapter 12 of Probabilistic Graphical models: Principles and Techniques by Daphne Koller and Nir

More information

CS 188: Artificial Intelligence Spring Announcements

CS 188: Artificial Intelligence Spring Announcements CS 188: Artificial Intelligence Spring 2011 Lecture 16: Bayes Nets IV Inference 3/28/2011 Pieter Abbeel UC Berkeley Many slides over this course adapted from Dan Klein, Stuart Russell, Andrew Moore Announcements

More information

Review: Directed Models (Bayes Nets)

Review: Directed Models (Bayes Nets) X Review: Directed Models (Bayes Nets) Lecture 3: Undirected Graphical Models Sam Roweis January 2, 24 Semantics: x y z if z d-separates x and y d-separation: z d-separates x from y if along every undirected

More information

Probabilistic Graphical Models: Representation and Inference

Probabilistic Graphical Models: Representation and Inference Probabilistic Graphical Models: Representation and Inference Aaron C. Courville Université de Montréal Note: Material for the slides is taken directly from a presentation prepared by Andrew Moore 1 Overview

More information

Lecture 4 October 18th

Lecture 4 October 18th Directed and undirected graphical models Fall 2017 Lecture 4 October 18th Lecturer: Guillaume Obozinski Scribe: In this lecture, we will assume that all random variables are discrete, to keep notations

More information

Pattern Recognition and Machine Learning

Pattern Recognition and Machine Learning Christopher M. Bishop Pattern Recognition and Machine Learning ÖSpri inger Contents Preface Mathematical notation Contents vii xi xiii 1 Introduction 1 1.1 Example: Polynomial Curve Fitting 4 1.2 Probability

More information

Probabilistic Graphical Models and Bayesian Networks. Artificial Intelligence Bert Huang Virginia Tech

Probabilistic Graphical Models and Bayesian Networks. Artificial Intelligence Bert Huang Virginia Tech Probabilistic Graphical Models and Bayesian Networks Artificial Intelligence Bert Huang Virginia Tech Concept Map for Segment Probabilistic Graphical Models Probabilistic Time Series Models Particle Filters

More information

17 : Markov Chain Monte Carlo

17 : Markov Chain Monte Carlo 10-708: Probabilistic Graphical Models, Spring 2015 17 : Markov Chain Monte Carlo Lecturer: Eric P. Xing Scribes: Heran Lin, Bin Deng, Yun Huang 1 Review of Monte Carlo Methods 1.1 Overview Monte Carlo

More information

Machine Learning 4771

Machine Learning 4771 Machine Learning 4771 Instructor: Tony Jebara Topic 11 Maximum Likelihood as Bayesian Inference Maximum A Posteriori Bayesian Gaussian Estimation Why Maximum Likelihood? So far, assumed max (log) likelihood

More information

Announcements. Inference. Mid-term. Inference by Enumeration. Reminder: Alarm Network. Introduction to Artificial Intelligence. V22.

Announcements. Inference. Mid-term. Inference by Enumeration. Reminder: Alarm Network. Introduction to Artificial Intelligence. V22. Introduction to Artificial Intelligence V22.0472-001 Fall 2009 Lecture 15: Bayes Nets 3 Midterms graded Assignment 2 graded Announcements Rob Fergus Dept of Computer Science, Courant Institute, NYU Slides

More information

Graphical Models and Independence Models

Graphical Models and Independence Models Graphical Models and Independence Models Yunshu Liu ASPITRG Research Group 2014-03-04 References: [1]. Steffen Lauritzen, Graphical Models, Oxford University Press, 1996 [2]. Christopher M. Bishop, Pattern

More information

Statistical Approaches to Learning and Discovery

Statistical Approaches to Learning and Discovery Statistical Approaches to Learning and Discovery Graphical Models Zoubin Ghahramani & Teddy Seidenfeld zoubin@cs.cmu.edu & teddy@stat.cmu.edu CALD / CS / Statistics / Philosophy Carnegie Mellon University

More information

ECE 6504: Advanced Topics in Machine Learning Probabilistic Graphical Models and Large-Scale Learning

ECE 6504: Advanced Topics in Machine Learning Probabilistic Graphical Models and Large-Scale Learning ECE 6504: Advanced Topics in Machine Learning Probabilistic Graphical Models and Large-Scale Learning Topics Summary of Class Advanced Topics Dhruv Batra Virginia Tech HW1 Grades Mean: 28.5/38 ~= 74.9%

More information

Intelligent Systems (AI-2)

Intelligent Systems (AI-2) Intelligent Systems (AI-2) Computer Science cpsc422, Lecture 19 Oct, 23, 2015 Slide Sources Raymond J. Mooney University of Texas at Austin D. Koller, Stanford CS - Probabilistic Graphical Models D. Page,

More information

σ(a) = a N (x; 0, 1 2 ) dx. σ(a) = Φ(a) =

σ(a) = a N (x; 0, 1 2 ) dx. σ(a) = Φ(a) = Until now we have always worked with likelihoods and prior distributions that were conjugate to each other, allowing the computation of the posterior distribution to be done in closed form. Unfortunately,

More information

STA 4273H: Statistical Machine Learning

STA 4273H: Statistical Machine Learning STA 4273H: Statistical Machine Learning Russ Salakhutdinov Department of Computer Science! Department of Statistical Sciences! rsalakhu@cs.toronto.edu! h0p://www.cs.utoronto.ca/~rsalakhu/ Lecture 7 Approximate

More information

Probabilistic Graphical Models

Probabilistic Graphical Models Probabilistic Graphical Models Lecture 9 Undirected Models CS/CNS/EE 155 Andreas Krause Announcements Homework 2 due next Wednesday (Nov 4) in class Start early!!! Project milestones due Monday (Nov 9)

More information

Undirected Graphical Models: Markov Random Fields

Undirected Graphical Models: Markov Random Fields Undirected Graphical Models: Markov Random Fields 40-956 Advanced Topics in AI: Probabilistic Graphical Models Sharif University of Technology Soleymani Spring 2015 Markov Random Field Structure: undirected

More information

Exact Inference I. Mark Peot. In this lecture we will look at issues associated with exact inference. = =

Exact Inference I. Mark Peot. In this lecture we will look at issues associated with exact inference. = = Exact Inference I Mark Peot In this lecture we will look at issues associated with exact inference 10 Queries The objective of probabilistic inference is to compute a joint distribution of a set of query

More information

Uncertainty and Bayesian Networks

Uncertainty and Bayesian Networks Uncertainty and Bayesian Networks Tutorial 3 Tutorial 3 1 Outline Uncertainty Probability Syntax and Semantics for Uncertainty Inference Independence and Bayes Rule Syntax and Semantics for Bayesian Networks

More information

Based on slides by Richard Zemel

Based on slides by Richard Zemel CSC 412/2506 Winter 2018 Probabilistic Learning and Reasoning Lecture 3: Directed Graphical Models and Latent Variables Based on slides by Richard Zemel Learning outcomes What aspects of a model can we

More information

Variable Elimination (VE) Barak Sternberg

Variable Elimination (VE) Barak Sternberg Variable Elimination (VE) Barak Sternberg Basic Ideas in VE Example 1: Let G be a Chain Bayesian Graph: X 1 X 2 X n 1 X n How would one compute P X n = k? Using the CPDs: P X 2 = x = x Val X1 P X 1 = x

More information

Recall from last time: Conditional probabilities. Lecture 2: Belief (Bayesian) networks. Bayes ball. Example (continued) Example: Inference problem

Recall from last time: Conditional probabilities. Lecture 2: Belief (Bayesian) networks. Bayes ball. Example (continued) Example: Inference problem Recall from last time: Conditional probabilities Our probabilistic models will compute and manipulate conditional probabilities. Given two random variables X, Y, we denote by Lecture 2: Belief (Bayesian)

More information

Chapter 16. Structured Probabilistic Models for Deep Learning

Chapter 16. Structured Probabilistic Models for Deep Learning Peng et al.: Deep Learning and Practice 1 Chapter 16 Structured Probabilistic Models for Deep Learning Peng et al.: Deep Learning and Practice 2 Structured Probabilistic Models way of using graphs to describe

More information

Inference in Bayesian Networks

Inference in Bayesian Networks Andrea Passerini passerini@disi.unitn.it Machine Learning Inference in graphical models Description Assume we have evidence e on the state of a subset of variables E in the model (i.e. Bayesian Network)

More information

Probabilistic Partial Evaluation: Exploiting rule structure in probabilistic inference

Probabilistic Partial Evaluation: Exploiting rule structure in probabilistic inference Probabilistic Partial Evaluation: Exploiting rule structure in probabilistic inference David Poole University of British Columbia 1 Overview Belief Networks Variable Elimination Algorithm Parent Contexts

More information

4 : Exact Inference: Variable Elimination

4 : Exact Inference: Variable Elimination 10-708: Probabilistic Graphical Models 10-708, Spring 2014 4 : Exact Inference: Variable Elimination Lecturer: Eric P. ing Scribes: Soumya Batra, Pradeep Dasigi, Manzil Zaheer 1 Probabilistic Inference

More information

Exact Inference: Clique Trees. Sargur Srihari

Exact Inference: Clique Trees. Sargur Srihari Exact Inference: Clique Trees Sargur srihari@cedar.buffalo.edu 1 Topics 1. Overview 2. Variable Elimination and Clique Trees 3. Message Passing: Sum-Product VE in a Clique Tree Clique-Tree Calibration

More information

Probabilistic Graphical Models

Probabilistic Graphical Models Probabilistic Graphical Models Brown University CSCI 2950-P, Spring 2013 Prof. Erik Sudderth Lecture 12: Gaussian Belief Propagation, State Space Models and Kalman Filters Guest Kalman Filter Lecture by

More information

Variable Elimination: Algorithm

Variable Elimination: Algorithm Variable Elimination: Algorithm Sargur srihari@cedar.buffalo.edu 1 Topics 1. Types of Inference Algorithms 2. Variable Elimination: the Basic ideas 3. Variable Elimination Sum-Product VE Algorithm Sum-Product

More information

EE562 ARTIFICIAL INTELLIGENCE FOR ENGINEERS

EE562 ARTIFICIAL INTELLIGENCE FOR ENGINEERS EE562 ARTIFICIAL INTELLIGENCE FOR ENGINEERS Lecture 16, 6/1/2005 University of Washington, Department of Electrical Engineering Spring 2005 Instructor: Professor Jeff A. Bilmes Uncertainty & Bayesian Networks

More information

Bayes Nets III: Inference

Bayes Nets III: Inference 1 Hal Daumé III (me@hal3.name) Bayes Nets III: Inference Hal Daumé III Computer Science University of Maryland me@hal3.name CS 421: Introduction to Artificial Intelligence 10 Apr 2012 Many slides courtesy

More information

Variable Elimination: Algorithm

Variable Elimination: Algorithm Variable Elimination: Algorithm Sargur srihari@cedar.buffalo.edu 1 Topics 1. Types of Inference Algorithms 2. Variable Elimination: the Basic ideas 3. Variable Elimination Sum-Product VE Algorithm Sum-Product

More information

Alternative Parameterizations of Markov Networks. Sargur Srihari

Alternative Parameterizations of Markov Networks. Sargur Srihari Alternative Parameterizations of Markov Networks Sargur srihari@cedar.buffalo.edu 1 Topics Three types of parameterization 1. Gibbs Parameterization 2. Factor Graphs 3. Log-linear Models Features (Ising,

More information

6.047 / Computational Biology: Genomes, Networks, Evolution Fall 2008

6.047 / Computational Biology: Genomes, Networks, Evolution Fall 2008 MIT OpenCourseWare http://ocw.mit.edu 6.047 / 6.878 Computational Biology: Genomes, Networks, Evolution Fall 2008 For information about citing these materials or our Terms of Use, visit: http://ocw.mit.edu/terms.

More information

Machine Learning Lecture 14

Machine Learning Lecture 14 Many slides adapted from B. Schiele, S. Roth, Z. Gharahmani Machine Learning Lecture 14 Undirected Graphical Models & Inference 23.06.2015 Bastian Leibe RWTH Aachen http://www.vision.rwth-aachen.de/ leibe@vision.rwth-aachen.de

More information

DEEP LEARNING CHAPTER 3 PROBABILITY & INFORMATION THEORY

DEEP LEARNING CHAPTER 3 PROBABILITY & INFORMATION THEORY DEEP LEARNING CHAPTER 3 PROBABILITY & INFORMATION THEORY OUTLINE 3.1 Why Probability? 3.2 Random Variables 3.3 Probability Distributions 3.4 Marginal Probability 3.5 Conditional Probability 3.6 The Chain

More information

Graphical Models. Lecture 3: Local Condi6onal Probability Distribu6ons. Andrew McCallum

Graphical Models. Lecture 3: Local Condi6onal Probability Distribu6ons. Andrew McCallum Graphical Models Lecture 3: Local Condi6onal Probability Distribu6ons Andrew McCallum mccallum@cs.umass.edu Thanks to Noah Smith and Carlos Guestrin for some slide materials. 1 Condi6onal Probability Distribu6ons

More information

UC Berkeley Department of Electrical Engineering and Computer Science Department of Statistics. EECS 281A / STAT 241A Statistical Learning Theory

UC Berkeley Department of Electrical Engineering and Computer Science Department of Statistics. EECS 281A / STAT 241A Statistical Learning Theory UC Berkeley Department of Electrical Engineering and Computer Science Department of Statistics EECS 281A / STAT 241A Statistical Learning Theory Solutions to Problem Set 2 Fall 2011 Issued: Wednesday,

More information

Massachusetts Institute of Technology Department of Electrical Engineering and Computer Science Algorithms For Inference Fall 2014

Massachusetts Institute of Technology Department of Electrical Engineering and Computer Science Algorithms For Inference Fall 2014 Massachusetts Institute of Technology Department of Electrical Engineering and Computer Science 6.438 Algorithms For Inference Fall 2014 Problem Set 3 Issued: Thursday, September 25, 2014 Due: Thursday,

More information

Lecture 17: May 29, 2002

Lecture 17: May 29, 2002 EE596 Pat. Recog. II: Introduction to Graphical Models University of Washington Spring 2000 Dept. of Electrical Engineering Lecture 17: May 29, 2002 Lecturer: Jeff ilmes Scribe: Kurt Partridge, Salvador

More information

Lecture 6: Graphical Models

Lecture 6: Graphical Models Lecture 6: Graphical Models Kai-Wei Chang CS @ Uniersity of Virginia kw@kwchang.net Some slides are adapted from Viek Skirmar s course on Structured Prediction 1 So far We discussed sequence labeling tasks:

More information

A graph contains a set of nodes (vertices) connected by links (edges or arcs)

A graph contains a set of nodes (vertices) connected by links (edges or arcs) BOLTZMANN MACHINES Generative Models Graphical Models A graph contains a set of nodes (vertices) connected by links (edges or arcs) In a probabilistic graphical model, each node represents a random variable,

More information

Undirected graphical models

Undirected graphical models Undirected graphical models Semantics of probabilistic models over undirected graphs Parameters of undirected models Example applications COMP-652 and ECSE-608, February 16, 2017 1 Undirected graphical

More information

Informatics 2D Reasoning and Agents Semester 2,

Informatics 2D Reasoning and Agents Semester 2, Informatics 2D Reasoning and Agents Semester 2, 2017 2018 Alex Lascarides alex@inf.ed.ac.uk Lecture 23 Probabilistic Reasoning with Bayesian Networks 15th March 2018 Informatics UoE Informatics 2D 1 Where

More information

Lecture : Probabilistic Machine Learning

Lecture : Probabilistic Machine Learning Lecture : Probabilistic Machine Learning Riashat Islam Reasoning and Learning Lab McGill University September 11, 2018 ML : Many Methods with Many Links Modelling Views of Machine Learning Machine Learning

More information

CS 2750: Machine Learning. Bayesian Networks. Prof. Adriana Kovashka University of Pittsburgh March 14, 2016

CS 2750: Machine Learning. Bayesian Networks. Prof. Adriana Kovashka University of Pittsburgh March 14, 2016 CS 2750: Machine Learning Bayesian Networks Prof. Adriana Kovashka University of Pittsburgh March 14, 2016 Plan for today and next week Today and next time: Bayesian networks (Bishop Sec. 8.1) Conditional

More information

CS281A/Stat241A Lecture 19

CS281A/Stat241A Lecture 19 CS281A/Stat241A Lecture 19 p. 1/4 CS281A/Stat241A Lecture 19 Junction Tree Algorithm Peter Bartlett CS281A/Stat241A Lecture 19 p. 2/4 Announcements My office hours: Tuesday Nov 3 (today), 1-2pm, in 723

More information

Graphical Models. Lecture 10: Variable Elimina:on, con:nued. Andrew McCallum

Graphical Models. Lecture 10: Variable Elimina:on, con:nued. Andrew McCallum Graphical Models Lecture 10: Variable Elimina:on, con:nued Andrew McCallum mccallum@cs.umass.edu Thanks to Noah Smith and Carlos Guestrin for some slide materials. 1 Last Time Probabilis:c inference is

More information

Markov Networks.

Markov Networks. Markov Networks www.biostat.wisc.edu/~dpage/cs760/ Goals for the lecture you should understand the following concepts Markov network syntax Markov network semantics Potential functions Partition function

More information

Learning Bayesian network : Given structure and completely observed data

Learning Bayesian network : Given structure and completely observed data Learning Bayesian network : Given structure and completely observed data Probabilistic Graphical Models Sharif University of Technology Spring 2017 Soleymani Learning problem Target: true distribution

More information

Nonparametric Bayesian Methods (Gaussian Processes)

Nonparametric Bayesian Methods (Gaussian Processes) [70240413 Statistical Machine Learning, Spring, 2015] Nonparametric Bayesian Methods (Gaussian Processes) Jun Zhu dcszj@mail.tsinghua.edu.cn http://bigml.cs.tsinghua.edu.cn/~jun State Key Lab of Intelligent

More information

CS 343: Artificial Intelligence

CS 343: Artificial Intelligence CS 343: Artificial Intelligence Bayes Nets: Sampling Prof. Scott Niekum The University of Texas at Austin [These slides based on those of Dan Klein and Pieter Abbeel for CS188 Intro to AI at UC Berkeley.

More information

Variational Inference. Sargur Srihari

Variational Inference. Sargur Srihari Variational Inference Sargur srihari@cedar.buffalo.edu 1 Plan of discussion We first describe inference with PGMs and the intractability of exact inference Then give a taxonomy of inference algorithms

More information

Introduction to Probabilistic Graphical Models

Introduction to Probabilistic Graphical Models Introduction to Probabilistic Graphical Models Sargur Srihari srihari@cedar.buffalo.edu 1 Topics 1. What are probabilistic graphical models (PGMs) 2. Use of PGMs Engineering and AI 3. Directionality in

More information

Bayesian Networks. instructor: Matteo Pozzi. x 1. x 2. x 3 x 4. x 5. x 6. x 7. x 8. x 9. Lec : Urban Systems Modeling

Bayesian Networks. instructor: Matteo Pozzi. x 1. x 2. x 3 x 4. x 5. x 6. x 7. x 8. x 9. Lec : Urban Systems Modeling 12735: Urban Systems Modeling Lec. 09 Bayesian Networks instructor: Matteo Pozzi x 1 x 2 x 3 x 4 x 5 x 6 x 7 x 8 x 9 1 outline example of applications how to shape a problem as a BN complexity of the inference

More information

Basic Sampling Methods

Basic Sampling Methods Basic Sampling Methods Sargur Srihari srihari@cedar.buffalo.edu 1 1. Motivation Topics Intractability in ML How sampling can help 2. Ancestral Sampling Using BNs 3. Transforming a Uniform Distribution

More information

Undirected Graphical Models

Undirected Graphical Models Outline Hong Chang Institute of Computing Technology, Chinese Academy of Sciences Machine Learning Methods (Fall 2012) Outline Outline I 1 Introduction 2 Properties Properties 3 Generative vs. Conditional

More information

CS 5522: Artificial Intelligence II

CS 5522: Artificial Intelligence II CS 5522: Artificial Intelligence II Bayes Nets: Independence Instructor: Alan Ritter Ohio State University [These slides were adapted from CS188 Intro to AI at UC Berkeley. All materials available at http://ai.berkeley.edu.]

More information

Introduction to Graphical Models

Introduction to Graphical Models Introduction to Graphical Models The 15 th Winter School of Statistical Physics POSCO International Center & POSTECH, Pohang 2018. 1. 9 (Tue.) Yung-Kyun Noh GENERALIZATION FOR PREDICTION 2 Probabilistic

More information

Inference in Hybrid Bayesian Networks with Mixtures of Truncated Exponentials

Inference in Hybrid Bayesian Networks with Mixtures of Truncated Exponentials In J. Vejnarova (ed.), Proceedings of 6th Workshop on Uncertainty Processing (WUPES-2003), 47--63, VSE-Oeconomica Publishers. Inference in Hybrid Bayesian Networks with Mixtures of Truncated Exponentials

More information

Conditional Independence and Factorization

Conditional Independence and Factorization Conditional Independence and Factorization Seungjin Choi Department of Computer Science and Engineering Pohang University of Science and Technology 77 Cheongam-ro, Nam-gu, Pohang 37673, Korea seungjin@postech.ac.kr

More information

What you need to know about Kalman Filters

What you need to know about Kalman Filters Readings: K&F: 4.5, 12.2, 12.3, 12.4, 18.1, 18.2, 18.3, 18.4 Switching Kalman Filter Dynamic Bayesian Networks Graphical Models 10708 Carlos Guestrin Carnegie Mellon University November 27 th, 2006 1 What

More information

Logistic Regression Review Fall 2012 Recitation. September 25, 2012 TA: Selen Uguroglu

Logistic Regression Review Fall 2012 Recitation. September 25, 2012 TA: Selen Uguroglu Logistic Regression Review 10-601 Fall 2012 Recitation September 25, 2012 TA: Selen Uguroglu!1 Outline Decision Theory Logistic regression Goal Loss function Inference Gradient Descent!2 Training Data

More information