Artificial Intelligence: Cognitive Agents
|
|
- Susanna Atkinson
- 6 years ago
- Views:
Transcription
1 Artificial Intelligence: Cognitive Agents AI, Uncertainty & Bayesian Networks / Kim, Byoung-Hee Biointelligence Laboratory Seoul National University
2 A Bayesian network is a graphical model for probabilistic relationships among a set of variables , SNU CSE Biointelligence Lab., 2
3 ACM Turing Award Nobel Prize in Computing 2011 Winner: Judea Pearl (UCLA) For fundamental contributions to artificial intelligence through the development of a calculus for probabilistic and causal reasoning Invention of Bayesian networks Pearl's accomplishments have redefined the term 'thinking machine over the past 30 years BN mimics the neural activities of the human brain, constantly exchanging messages without benefit of a supervisor , SNU CSE Biointelligence Lab., 3
4 Notes Related chapters in the textbook (AIMA 3 rd ed. by Russell and Norvig) Ch. 13 Quantifying Uncertainty Ch. 14 Probabilistic Reasoning (14.1~14.2) Ch. 20 Learning Probabilistic Models (20.1~20.2) On the reference A Tutorial on Learning with Bayesian Networks by David Heckerman It is very technical but covers insights and comprehensive backgrounds on Bayesian networks This lecture covers the Introduction section This lecture is an easier introductory tutorial Both contents in the textbook and Heckerman s tutorial is fairly mathematical This lecture covers basic concepts and tools to understand Bayesian networks , SNU CSE Biointelligence Lab., 4
5 Contents Bayesian Networks: Introduction Motivating example Decomposing a joint distribution of variables d-separation A mini Turing test in causal conversation Correlation & causation AI & Uncertainty Bayesian Networks in Detail d-separation: revisited & details Probability & Bayesian Inference in & learning Bayesian networks BN as AI tools and advantages , SNU CSE Biointelligence Lab., 5
6 Bayesian Networks: Introduction
7 Causality, Dependency From correlation to causality 정성적방법 Slide from K. Mohan & J. Pearl, UAI 12 Tutorial on Graphical Models for Causal Inference 정량적방법 Granger causality index Google 의 CausalImpact R package for causal inference in time series Official posting: 소개기사 ( 영문 ): , SNU CSE Biointelligence Lab., 7
8 Slide from K. Mohan & J. Pearl, UAI 12 Tutorial on Graphical Models for Causal Inference Assuming binary states for all the variables Ex) Season: dry or rainy Ex) Sprinkler: ON or OFF , SNU CSE Biointelligence Lab., 8
9 Slide from K. Mohan & J. Pearl, UAI 12 Tutorial on Graphical Models for Causal Inference , SNU CSE Biointelligence Lab., 9
10 Slide from K. Mohan & J. Pearl, UAI 12 Tutorial on Graphical Models for Causal Inference (directed acyclic graph) , SNU CSE Biointelligence Lab., 10
11 Slide from K. Mohan & J. Pearl, UAI 12 Tutorial on Graphical Models for Causal Inference , SNU CSE Biointelligence Lab., 11
12 Slide from K. Mohan & J. Pearl, UAI 12 Tutorial on Graphical Models for Causal Inference , SNU CSE Biointelligence Lab., 12
13 Slide from K. Mohan & J. Pearl, UAI 12 Tutorial on Graphical Models for Causal Inference , SNU CSE Biointelligence Lab., 13
14 Slide from K. Mohan & J. Pearl, UAI 12 Tutorial on Graphical Models for Causal Inference , SNU CSE Biointelligence Lab., 14
15 Slide from K. Mohan & J. Pearl, UAI 12 Tutorial on Graphical Models for Causal Inference , SNU CSE Biointelligence Lab., 15
16 Slide from K. Mohan & J. Pearl, UAI 12 Tutorial on Graphical Models for Causal Inference , SNU CSE Biointelligence Lab., 16
17 Slide from K. Mohan & J. Pearl, UAI 12 Tutorial on Graphical Models for Causal Inference , SNU CSE Biointelligence Lab., 17
18 Slide from K. Mohan & J. Pearl, UAI 12 Tutorial on Graphical Models for Causal Inference , SNU CSE Biointelligence Lab., 18
19 Slide from K. Mohan & J. Pearl, UAI 12 Tutorial on Graphical Models for Causal Inference , SNU CSE Biointelligence Lab., 19
20 Slide from K. Mohan & J. Pearl, UAI 12 Tutorial on Graphical Models for Causal Inference , SNU CSE Biointelligence Lab., 20
21 Slide from K. Mohan & J. Pearl, UAI 12 Tutorial on Graphical Models for Causal Inference , SNU CSE Biointelligence Lab., 21
22 Slide from K. Mohan & J. Pearl, UAI 12 Tutorial on Graphical Models for Causal Inference , SNU CSE Biointelligence Lab., 22
23 Slide from K. Mohan & J. Pearl, UAI 12 Tutorial on Graphical Models for Causal Inference , SNU CSE Biointelligence Lab., 23
24 Slide from K. Mohan & J. Pearl, UAI 12 Tutorial on Graphical Models for Causal Inference , SNU CSE Biointelligence Lab., 24
25 Correlation & Causation Correlation does not imply causation A chart that, correlates the number of pirates with global temperature. The two variables are correlated, but one does not imply the other a correlation between ice cream consumption and crime, but shows that the actual cause is temperature , SNU CSE Biointelligence Lab., 25
26 Example Designing a Bayesian Network My own design of conditional probability tables SEASON DRY 0.6 SPRINKLER SEASON DRY RAINY DRY RAINY RAINY 0.4 WET RAIN SEASON DRY RAINY YES NO SPRINKLER ON OFF RAIN YES NO YES NO YES NO SLIPPERY WET YES NO YES NO , SNU CSE Biointelligence Lab., 26
27 Example Designing a Bayesian Network Tool: GeNIe ( GeNIe (Graphical Network Interface) is the graphical interface to SMILE, a fully portable Bayesian inference engine in C++ Inference based on the designed Bayesian Network Q1 A , SNU CSE Biointelligence Lab., 27
28 AI & Uncertainty
29 Probability Probability plays a central role in modern pattern recognition. The main tool to deal uncertainties All of the probabilistic inference and learning amount to repeated application of the sum rule and the product rule Random Variables: variables + probability , SNU CSE Biointelligence Lab., 29
30 Artificial Intelligence (AI) The objective of AI is to build intelligent computers We want intelligent, adaptive, robust behavior cat car Often hand programming not possible. Solution? Get the computer to program itself, by showing it examples of the behavior we want! This is the learning approach to AI , SNU CSE Biointelligence Lab., 30
31 Artificial Intelligence (AI) (Traditional) AI Knowledge & reasoning; work with facts/assertions; develop rules of logical inference Planning: work with applicability/effects of actions; develop searches for actions which achieve goals/avert disasters. Expert systems: develop by hand a set of rules for examining inputs, updating internal states and generating outputs , SNU CSE Biointelligence Lab., 31
32 Artificial Intelligence (AI) Probabilistic AI emphasis on noisy measurements, approximation in hard cases, learning, algorithmic issues. The power of learning Automatic system building old expert systems needed hand coding of knowledge and of output semantics learning automatically constructs rules and supports all types of queries Probabilistic databases traditional DB technology cannot answer queries about items that were never loaded into the dataset UAI models are like probabilistic databases , SNU CSE Biointelligence Lab., 32
33 Uncertainty and Artificial Intelligence (UAI) Probabilistic methods can be used to: make decisions given partial information about the world account for noisy sensors or actuators explain phenomena not part of our models describe inherently stochastic behavior in the world , SNU CSE Biointelligence Lab., 33
34 Other Names for UAI Machine learning (ML), data mining, applied statistics, adaptive (stochastic) signal processing, probabilistic planning/reasoning... Some differences: Data mining almost always uses large data sets, statistics almost always small ones Data mining, planning, decision theory often have no internal parameters to be learned Statistics often has no algorithm to run! ML/UAI algorithms are rarely online and rarely scale to huge data (changing now) , SNU CSE Biointelligence Lab., 34
35 Learning is most useful Learning in AI when the structure of the task is not well understood but can be characterized by a dataset with strong statistical regularity Also useful in adaptive or dynamic situations when the task (or its parameters) are constantly changing Currently, these are challenging topics of machine learning and data mining research , SNU CSE Biointelligence Lab., 35
36 Probabilistic AI Let inputs=x, correct answers=y, outputs of our machine=z Learning: estimation of p(x, Y) The central object of interest is the joint distribution The main difficulty is compactly representing it and robustly learning its shape given noisy samples , SNU CSE Biointelligence Lab., 36
37 Probabilistic Graphical Models (PGMs) Probabilistic graphical models represent large joint distributions compactly using a set of local relationships specified by a graph Each random variable in our model corresponds to a graph node , SNU CSE Biointelligence Lab., 37
38 Probabilistic Graphical Models (PGMs) There are useful properties in using probabilistic graphical models A simple way to visualize the structure of a probabilistic model Insights into the properties of the model Complex computations (for inference and learning) can be expressed in terms of graphical manipulations underlying mathematical expressions , SNU CSE Biointelligence Lab., 38
39 Directed graph vs. undirected graph Both (probabilistic) graphical models Specify a factorization (how to express the joint distribution) Define a set of conditional independence properties Parent - child Local conditional distribution Maximal clique Potential function Bayesian Networks (BN) Markov Random Field (MRF) , SNU CSE Biointelligence Lab., 39
40 Bayesian Networks in Detail
41 (DAG) , SNU CSE Biointelligence Lab., 41
42 Designing a Bayesian Network Model TakeHeart II: Decision support system for clinical cardiovascular risk assessment , SNU CSE Biointelligence Lab., 42
43 Inference in a Bayesian Network Model Given an assignment of a subset of variables (evidence) in a BN, estimate the posterior distribution over another subset of unobserved variables of interest. Inferences viewed as message passing along the network , SNU CSE Biointelligence Lab., 43
44 Bayesian Networks The joint distribution defined by a graph is given by the product of a conditional distribution of each node conditioned on their parent nodes. p(x) = K k = 1 p( x Pa( k x k )) (PP(x k ) denotes the set of parents of x k ) ex) p x 1, x 2,, x 7 = * Without given DAG structure, usual chain rule can be applied to get the joint distribution. But computational cost is much higher , SNU CSE Biointelligence Lab., 44
45 Bayes Theorem Likelihood py ( X) = p( X Y) py ( ) px ( ) Prior Posterior Normalizing constant px ( ) = px ( Y) py ( ) Y posterior likelihood prior , SNU CSE Biointelligence Lab., 45
46 Bayes Theorem Figure from Figure 1. in (Adams, et all, 2013) obtained from , SNU CSE Biointelligence Lab., 46
47 Likelihood: Frequentist Bayesian Probabilities -Frequentist vs. Bayesian w: a fixed parameter determined by estimator Maximum likelihood: Error function = log p( D w) Error bars: Obtained by the distribution of possible data sets Bootstrap Cross-validation Bayesian p( D w) p( w D) = p( D w) p( w) p( D) Thomas Bayes a probability distribution w: the uncertainty in the parameters Prior knowledge Noninformative (uniform) prior, Laplace correction in estimating priors Monte Carlo methods, variational Bayes, EP D (See an article WHERE Do PROBABILITIES COME FROM? on page 491 in the textbook (Russell and Norvig, 2010) for more discussion) , SNU CSE Biointelligence Lab., 47
48 Conditional Independence Conditional independence simplifies both the structure of a model and the computations An important feature of graphical models is that conditional independence properties of the joint distribution can be read directly from the graph without having to perform any analytical manipulations The general framework for this is called d-separation , SNU CSE Biointelligence Lab., 48
49 Three example graphs 1 st case None of the variables are observed Node c is tail-to-tail The variable c is observed The conditioned node blocks the path from a to b, causes a and b to become (conditionally) independent , SNU CSE Biointelligence Lab., 49
50 Three example graphs 2 nd case None of the variables are observed Node c is head-to-tail The variable c is observed The conditioned node blocks the path from a to b, causes a and b to become (conditionally) independent , SNU CSE Biointelligence Lab., 50
51 Three example graphs 3 rd case None of the variables are observed The variable c is observed Node c is head-to-head When node c is unobserved, it blocks the path and the variables a and b are independent. Conditioning on c unblocks the path and render a and b dependent , SNU CSE Biointelligence Lab., 51
52 Three example graphs - Fuel gauge example B Battery, F-fuel, G-electric fuel gauge (rather unreliable fuel gauge) Checking the fuel gauge Checking the battery also has the meaning? ( Makes it more likely ) Makes it less likely than observation of fuel gauge only. (explaining away) , SNU CSE Biointelligence Lab., 52
53 d-separation Tail-to-tail node or head-to-tail node Unless it is observed in which case it blocks a path, the path is unblocked. Head-to-head node Blocks a path if is unobserved, but on the node, and/or at least one of its descendants, is observed the path becomes unblocked. d-separation? All paths are blocked. The joint distribution will satisfy conditional independence w.r.t. concerned variables , SNU CSE Biointelligence Lab., 53
54 d-separation (a) a is dependent to b given c Head-to-head node e is unblocked, because a descendant c is in the conditioning set. Tail-to-tail node f is unblocked (b) a is independent to b given f Head-to-head node e is blocked Tail-to-tail node f is blocked , SNU CSE Biointelligence Lab., 54
55 d-separation Another example of conditional independence and d-separation: i.i.d. (independent identically distributed) data Problem: finding posterior dist. for the mean of a univariate Gaussian dist. Every path is blocked and so the observations D={x 1,,x N } are independent given (independent) (The observations are in general no longer independent!) , SNU CSE Biointelligence Lab., 55
56 d-separation Naïve Bayes model Key assumption: conditioned on the class z, the distribution of the input variables x 1,, x D are independent. Input {x 1,,x N } with their class labels, then we can fit the naïve Bayes model to the training data using maximum likelihood assuming that the data are drawn independently from the model , SNU CSE Biointelligence Lab., 56
57 d-separation Markov blanket or Markov boundary When dealing with the conditional distribution of x i, consider the minimal set of nodes that isolates x i from the rest of the graph. The set of nodes comprising parents, children, co-parents is called the Markov blanket. parents Co-parents children , SNU CSE Biointelligence Lab., 57
58 Probability Distributions Discrete variables Beta, Bernoulli, binomial Dirichlet, multinomial Continuous variables Normal (Gaussian) Student-t beta Dirichlet Exponential family & conjugacy Many probability densities on x can be represented as the same form T { } p( x η) = h( x) g( η)exp η ux ( ) binomial Gaussian There are conjugate family of density functions having the same form of density functions Beta & binomial F beta Dirichlet Dirichlet & multinomial Normal & Normal x binomial multinomial , SNU CSE Biointelligence Lab., 58
59 Inference in Graphical Models Inference in graphical models Given evidences (some nodes are clamped to observed values) Wish to compute the posterior distributions of other nodes Inference algorithms in graphical structures Main idea: propagation of local messages Exact inference Sum-product algorithm, max-product algorithm, junction tree A algorithm Approximate inference Loopy belief propagation + message passing schedule Variational methods, sampling methods (Monte Carlo methods) B C D E ABD BCD CDE , SNU CSE Biointelligence Lab., 59
60 Parameters Learning Parameters of Bayesian Networks probabilities in conditional probability tables (CPTs) for all the variables in the network Learning parameters SEASON Assuming that the structure is fixed, i.e. designed or learned. We need data, i.e. observed instances Estimation based on relative frequencies from data + belief Example: coin toss. Estimation of heads in various ways RAIN SEASON DRY RAINY DRY? YES?? RAINY? NO?? 1 2 The principle of indifference: head and tail are equally probable If we tossed a coin 10,000 times and it landed heads 3373 times, we would estimate the probability of heads to be about.3373 P heeee = , SNU CSE Biointelligence Lab., 60
61 Learning Parameters of Bayesian Networks Learning parameters (continued) Estimation based on relative frequencies from data + belief Example: A-match soccer game between Korea and Japan. How, do you think, is it probable that Korean would win? A: 0.85 (Korean), B: 0.3 (Japanese) 3 This probability is not a ratio, and it is not a relative frequency because the game cannot be repeated many times under the exact same conditions Degree of belief or subjective probability Usual method Estimate the probability distribution of a variable X based on a relative frequency and belief concerning a relative frequency , SNU CSE Biointelligence Lab., 61
62 Learning Parameters of Bayesian Networks Simple counting solution (Bayesian point of view) Parameter estimation of a single node Assume local parameter independence For a binary variable (for example, a coin toss) prior: Beta distribution - Beta(a,b) after we have observed m heads and N-m tails posterior - Beta(a+m,b+N-m) and P X = heee = (a+m) N (conjugacy of Beta and Binomial distributions) beta binomial , SNU CSE Biointelligence Lab., 62
63 Learning Parameters of Bayesian Networks Simple counting solution (Bayesian point of view) For a multinomial variable (for example, a dice toss) prior: Dirichlet distribution Dirichlet(a 1,a 2,, a d ) P X = k = a k N N = a k Observing state i: Dirichlet(a 1,,a i +1,, a d ) (conjugacy of Dirichlet and Multinomial distributions) For an entire network We simply iterate over its nodes In the case of incomplete data In real data, many of the variable values may be incorrect or missing Usual approximating solution is given by Gibbs sampling or EM (expectation maximization) technique , SNU CSE Biointelligence Lab., 63
64 Smoothing Another viewpoint Learning Parameters of Bayesian Networks Laplace smoothing or additive smoothing given observed counts for d states of a variable X = (x 1, x 2, x d ) P X = k = x k + α N + αd i = 1,, d, (α = α 1 = α 2 = α d ) From a Bayesian point of view, this corresponds to the expected value of the posterior distribution, using a symmetric Dirichlet distribution with parameter α as a prior. Additive smoothing is commonly a component of naive Bayes classifiers , SNU CSE Biointelligence Lab., 64
65 Learning the Graph Structure Learning the graph structure itself from data requires A space of possible structures A measure that can be used to score each structure From a Bayesian viewpoint : score for each model Tough points Marginalization over latent variables => challenging computational problem Exploring the space of structures can also be problematic The # of different graph structures grows exponentially with the # of nodes Usually we resort to heuristics Local score based, global score based, conditional independence test based, , SNU CSE Biointelligence Lab., 65
66 Bayesian Networks as Tools for AI Learning Extracting and encoding knowledge from data Knowledge is represented in Probabilistic relationship among variables Causal relationship Network of variables Common framework for machine learning models Supervised and unsupervised learning Knowledge Representation & Reasoning Bayesian networks can be constructed from prior knowledge alone Constructed model can be used for reasoning based on probabilistic inference methods Expert System Uncertain expert knowledge can be encoded into a Bayesian network DAG in a Bayesian network is hand-constructed by domain experts Then the conditional probabilities were assessed by the expert, learned from data, or obtained using a combination of both techniques. Bayesian network-based expert systems are popular Planning In some different form, known as decision graphs or influence diagrams We don t cover about this direction , SNU CSE Biointelligence Lab., 66
67 Advantages of Bayesian Networks for Data Analysis Ability to handle missing data Because the model encodes dependencies among all variables Learning causal relationships Can be used to gain understanding about a problem domain Can be used to predict the consequences of intervention Having both causal and probabilistic semantics It is an ideal representation for combining prior knowledge (which comes in causal form) and data Efficient and principled approach for avoiding the overfitting of data By Bayesian statistical methods in conjunction with Bayesian networks (summary from the abstract of D. Heckerman s Tutorial on BN) (Read Introduction section for detailed explanations) , SNU CSE Biointelligence Lab., 67
68 References K. Mohan & J. Pearl, UAI 12 Tutorial on Graphical Models for Causal Inference S. Roweis, MLSS 05 Lecture on Probabilistic Graphical Models Chapter 1, Chapter 2, Chapter 8 (Graphical Models), in Pattern Recognition and Machine Learning by C.M. Bishop, David Heckerman, A Tutorial on Learning with Bayesian Networks. R.E. Neapolitan, Learning Bayesian Networks, Pearson Prentice Hall, , SNU CSE Biointelligence Lab., 68
69 More Textbooks and Courses : Probabilistic Graphical Models by D. Koller , SNU CSE Biointelligence Lab., 69
70 APPENDIX , SNU CSE Biointelligence Lab., 70
71 Learning Parameters of Bayesian Networks F The probability distribution of F represents our belief concerning the relative frequency with which X equals k. X Case 1: X is a binary variable F: beta distribution, X: Bernoulli or binomial distribution Ex) F ~ Beta(a,b), then P X = 1 = a N (N = a + b) Case 2: X is a multinomial variable F: Dirichlet distribution, X: multinomial distribution Ex) F ~ Dirichlet(a 1,a 2,, a d ), then P X = k = a k N N = a k Laplace smoothing or additive smoothing given observed frequencies X = (x 1, x 2, x d ) P X = k = x k + α i = 1,, d, (α = α N + αd 1 = α 2 = α d ) , SNU CSE Biointelligence Lab., 71
72 Graphical interpretation of Given structure: Bayes theorem pxy (, ) = pxpy ( ) ( x) We observe the value of y Goal: infer the posterior distribution over x, px ( y) Marginal distribution px ( ) : a prior over the latent variable x py ( ) We can evaluate the marginal distribution py ( ) = py ( x') px ( ') x' (b) By Bayes theorem we can calculate (a) px ( y) = py ( xpx ) ( ) py ( ) (c) , SNU CSE Biointelligence Lab., 72
73 d-separation Directed factorization Filtering whether can be expressed in terms of the factorization implied by the graph? If we present to the filter the set of all possible distributions p(x) over the set of variables X, then the subset of distributions that are passed by the filter will be denoted DF (Directed Factorization) Fully connected graph: The set DF will contain all possible distributions Fully disconnected graph: The joint distributions which factorize into the product of the marginal distributions over the variables only , SNU CSE Biointelligence Lab., 73
74 Gaussian distribution N 1 1 ( x µσ, ) = exp ( x µ ) (2 πσ ) 2σ /2 2 Multivariate Gaussian N x μ Σ = x μ Σ x μ (2 π ) Σ 2 T 1 (, ) exp ( ) ( ) D/2 1/ , SNU CSE Biointelligence Lab., 74
Bayesian Networks Practice
Bayesian Networks Practice Part 2 2016-03-17 Byoung-Hee Kim, Seong-Ho Son Biointelligence Lab, CSE, Seoul National University Agenda Probabilistic Inference in Bayesian networks Probability basics D-searation
More informationBayesian Networks BY: MOHAMAD ALSABBAGH
Bayesian Networks BY: MOHAMAD ALSABBAGH Outlines Introduction Bayes Rule Bayesian Networks (BN) Representation Size of a Bayesian Network Inference via BN BN Learning Dynamic BN Introduction Conditional
More informationCS 2750: Machine Learning. Bayesian Networks. Prof. Adriana Kovashka University of Pittsburgh March 14, 2016
CS 2750: Machine Learning Bayesian Networks Prof. Adriana Kovashka University of Pittsburgh March 14, 2016 Plan for today and next week Today and next time: Bayesian networks (Bishop Sec. 8.1) Conditional
More informationp L yi z n m x N n xi
y i z n x n N x i Overview Directed and undirected graphs Conditional independence Exact inference Latent variables and EM Variational inference Books statistical perspective Graphical Models, S. Lauritzen
More informationLecture 6: Graphical Models: Learning
Lecture 6: Graphical Models: Learning 4F13: Machine Learning Zoubin Ghahramani and Carl Edward Rasmussen Department of Engineering, University of Cambridge February 3rd, 2010 Ghahramani & Rasmussen (CUED)
More informationChris Bishop s PRML Ch. 8: Graphical Models
Chris Bishop s PRML Ch. 8: Graphical Models January 24, 2008 Introduction Visualize the structure of a probabilistic model Design and motivate new models Insights into the model s properties, in particular
More informationMachine Learning Summer School
Machine Learning Summer School Lecture 3: Learning parameters and structure Zoubin Ghahramani zoubin@eng.cam.ac.uk http://learning.eng.cam.ac.uk/zoubin/ Department of Engineering University of Cambridge,
More informationBayesian Models in Machine Learning
Bayesian Models in Machine Learning Lukáš Burget Escuela de Ciencias Informáticas 2017 Buenos Aires, July 24-29 2017 Frequentist vs. Bayesian Frequentist point of view: Probability is the frequency of
More informationIntroduction to Probabilistic Graphical Models
Introduction to Probabilistic Graphical Models Kyu-Baek Hwang and Byoung-Tak Zhang Biointelligence Lab School of Computer Science and Engineering Seoul National University Seoul 151-742 Korea E-mail: kbhwang@bi.snu.ac.kr
More informationECE521 Tutorial 11. Topic Review. ECE521 Winter Credits to Alireza Makhzani, Alex Schwing, Rich Zemel and TAs for slides. ECE521 Tutorial 11 / 4
ECE52 Tutorial Topic Review ECE52 Winter 206 Credits to Alireza Makhzani, Alex Schwing, Rich Zemel and TAs for slides ECE52 Tutorial ECE52 Winter 206 Credits to Alireza / 4 Outline K-means, PCA 2 Bayesian
More informationBayesian Networks Practice
ayesian Networks Practice Part 2 2016-03-17 young-hee Kim Seong-Ho Son iointelligence ab CSE Seoul National University Agenda Probabilistic Inference in ayesian networks Probability basics D-searation
More informationDirected Graphical Models
CS 2750: Machine Learning Directed Graphical Models Prof. Adriana Kovashka University of Pittsburgh March 28, 2017 Graphical Models If no assumption of independence is made, must estimate an exponential
More informationMachine Learning Lecture 14
Many slides adapted from B. Schiele, S. Roth, Z. Gharahmani Machine Learning Lecture 14 Undirected Graphical Models & Inference 23.06.2015 Bastian Leibe RWTH Aachen http://www.vision.rwth-aachen.de/ leibe@vision.rwth-aachen.de
More informationGraphical Models and Kernel Methods
Graphical Models and Kernel Methods Jerry Zhu Department of Computer Sciences University of Wisconsin Madison, USA MLSS June 17, 2014 1 / 123 Outline Graphical Models Probabilistic Inference Directed vs.
More informationProbabilistic Graphical Models (I)
Probabilistic Graphical Models (I) Hongxin Zhang zhx@cad.zju.edu.cn State Key Lab of CAD&CG, ZJU 2015-03-31 Probabilistic Graphical Models Modeling many real-world problems => a large number of random
More informationBayesian Networks: Construction, Inference, Learning and Causal Interpretation. Volker Tresp Summer 2016
Bayesian Networks: Construction, Inference, Learning and Causal Interpretation Volker Tresp Summer 2016 1 Introduction So far we were mostly concerned with supervised learning: we predicted one or several
More informationPMR Learning as Inference
Outline PMR Learning as Inference Probabilistic Modelling and Reasoning Amos Storkey Modelling 2 The Exponential Family 3 Bayesian Sets School of Informatics, University of Edinburgh Amos Storkey PMR Learning
More informationLecture 1: Probabilistic Graphical Models. Sam Roweis. Monday July 24, 2006 Machine Learning Summer School, Taiwan
Lecture 1: Probabilistic Graphical Models Sam Roweis Monday July 24, 2006 Machine Learning Summer School, Taiwan Building Intelligent Computers We want intelligent, adaptive, robust behaviour in computers.
More informationPart I. C. M. Bishop PATTERN RECOGNITION AND MACHINE LEARNING CHAPTER 8: GRAPHICAL MODELS
Part I C. M. Bishop PATTERN RECOGNITION AND MACHINE LEARNING CHAPTER 8: GRAPHICAL MODELS Probabilistic Graphical Models Graphical representation of a probabilistic model Each variable corresponds to a
More informationA graph contains a set of nodes (vertices) connected by links (edges or arcs)
BOLTZMANN MACHINES Generative Models Graphical Models A graph contains a set of nodes (vertices) connected by links (edges or arcs) In a probabilistic graphical model, each node represents a random variable,
More informationAn Introduction to Bayesian Machine Learning
1 An Introduction to Bayesian Machine Learning José Miguel Hernández-Lobato Department of Engineering, Cambridge University April 8, 2013 2 What is Machine Learning? The design of computational systems
More informationConditional Independence
Conditional Independence Sargur Srihari srihari@cedar.buffalo.edu 1 Conditional Independence Topics 1. What is Conditional Independence? Factorization of probability distribution into marginals 2. Why
More informationBayesian Networks: Construction, Inference, Learning and Causal Interpretation. Volker Tresp Summer 2014
Bayesian Networks: Construction, Inference, Learning and Causal Interpretation Volker Tresp Summer 2014 1 Introduction So far we were mostly concerned with supervised learning: we predicted one or several
More informationArtificial Intelligence Bayes Nets: Independence
Artificial Intelligence Bayes Nets: Independence Instructors: David Suter and Qince Li Course Delivered @ Harbin Institute of Technology [Many slides adapted from those created by Dan Klein and Pieter
More informationProbabilistic Graphical Networks: Definitions and Basic Results
This document gives a cursory overview of Probabilistic Graphical Networks. The material has been gleaned from different sources. I make no claim to original authorship of this material. Bayesian Graphical
More information{ p if x = 1 1 p if x = 0
Discrete random variables Probability mass function Given a discrete random variable X taking values in X = {v 1,..., v m }, its probability mass function P : X [0, 1] is defined as: P (v i ) = Pr[X =
More informationSTA 4273H: Statistical Machine Learning
STA 4273H: Statistical Machine Learning Russ Salakhutdinov Department of Statistics! rsalakhu@utstat.toronto.edu! http://www.utstat.utoronto.ca/~rsalakhu/ Sidney Smith Hall, Room 6002 Lecture 3 Linear
More informationIntroduction to Bayesian Learning
Course Information Introduction Introduction to Bayesian Learning Davide Bacciu Dipartimento di Informatica Università di Pisa bacciu@di.unipi.it Apprendimento Automatico: Fondamenti - A.A. 2016/2017 Outline
More informationGraphical Models - Part I
Graphical Models - Part I Oliver Schulte - CMPT 726 Bishop PRML Ch. 8, some slides from Russell and Norvig AIMA2e Outline Probabilistic Models Bayesian Networks Markov Random Fields Inference Outline Probabilistic
More informationGrundlagen der Künstlichen Intelligenz
Grundlagen der Künstlichen Intelligenz Uncertainty & Probabilities & Bandits Daniel Hennes 16.11.2017 (WS 2017/18) University Stuttgart - IPVS - Machine Learning & Robotics 1 Today Uncertainty Probability
More informationBayesian Networks. instructor: Matteo Pozzi. x 1. x 2. x 3 x 4. x 5. x 6. x 7. x 8. x 9. Lec : Urban Systems Modeling
12735: Urban Systems Modeling Lec. 09 Bayesian Networks instructor: Matteo Pozzi x 1 x 2 x 3 x 4 x 5 x 6 x 7 x 8 x 9 1 outline example of applications how to shape a problem as a BN complexity of the inference
More informationBayesian Machine Learning
Bayesian Machine Learning Andrew Gordon Wilson ORIE 6741 Lecture 4 Occam s Razor, Model Construction, and Directed Graphical Models https://people.orie.cornell.edu/andrew/orie6741 Cornell University September
More informationCS 5522: Artificial Intelligence II
CS 5522: Artificial Intelligence II Bayes Nets: Independence Instructor: Alan Ritter Ohio State University [These slides were adapted from CS188 Intro to AI at UC Berkeley. All materials available at http://ai.berkeley.edu.]
More informationCPSC 540: Machine Learning
CPSC 540: Machine Learning Undirected Graphical Models Mark Schmidt University of British Columbia Winter 2016 Admin Assignment 3: 2 late days to hand it in today, Thursday is final day. Assignment 4:
More informationCourse Introduction. Probabilistic Modelling and Reasoning. Relationships between courses. Dealing with Uncertainty. Chris Williams.
Course Introduction Probabilistic Modelling and Reasoning Chris Williams School of Informatics, University of Edinburgh September 2008 Welcome Administration Handout Books Assignments Tutorials Course
More informationLearning in Bayesian Networks
Learning in Bayesian Networks Florian Markowetz Max-Planck-Institute for Molecular Genetics Computational Molecular Biology Berlin Berlin: 20.06.2002 1 Overview 1. Bayesian Networks Stochastic Networks
More informationPattern Recognition and Machine Learning
Christopher M. Bishop Pattern Recognition and Machine Learning ÖSpri inger Contents Preface Mathematical notation Contents vii xi xiii 1 Introduction 1 1.1 Example: Polynomial Curve Fitting 4 1.2 Probability
More informationProbabilistic Graphical Models for Image Analysis - Lecture 1
Probabilistic Graphical Models for Image Analysis - Lecture 1 Alexey Gronskiy, Stefan Bauer 21 September 2018 Max Planck ETH Center for Learning Systems Overview 1. Motivation - Why Graphical Models 2.
More information13: Variational inference II
10-708: Probabilistic Graphical Models, Spring 2015 13: Variational inference II Lecturer: Eric P. Xing Scribes: Ronghuo Zheng, Zhiting Hu, Yuntian Deng 1 Introduction We started to talk about variational
More information6.047 / Computational Biology: Genomes, Networks, Evolution Fall 2008
MIT OpenCourseWare http://ocw.mit.edu 6.047 / 6.878 Computational Biology: Genomes, Networks, Evolution Fall 2008 For information about citing these materials or our Terms of Use, visit: http://ocw.mit.edu/terms.
More informationPATTERN RECOGNITION AND MACHINE LEARNING CHAPTER 2: PROBABILITY DISTRIBUTIONS
PATTERN RECOGNITION AND MACHINE LEARNING CHAPTER 2: PROBABILITY DISTRIBUTIONS Parametric Distributions Basic building blocks: Need to determine given Representation: or? Recall Curve Fitting Binary Variables
More informationChapter 16. Structured Probabilistic Models for Deep Learning
Peng et al.: Deep Learning and Practice 1 Chapter 16 Structured Probabilistic Models for Deep Learning Peng et al.: Deep Learning and Practice 2 Structured Probabilistic Models way of using graphs to describe
More informationProbabilistic Graphical Models. Guest Lecture by Narges Razavian Machine Learning Class April
Probabilistic Graphical Models Guest Lecture by Narges Razavian Machine Learning Class April 14 2017 Today What is probabilistic graphical model and why it is useful? Bayesian Networks Basic Inference
More informationProbabilistic Graphical Models: Representation and Inference
Probabilistic Graphical Models: Representation and Inference Aaron C. Courville Université de Montréal Note: Material for the slides is taken directly from a presentation prepared by Andrew Moore 1 Overview
More informationCSC321 Lecture 18: Learning Probabilistic Models
CSC321 Lecture 18: Learning Probabilistic Models Roger Grosse Roger Grosse CSC321 Lecture 18: Learning Probabilistic Models 1 / 25 Overview So far in this course: mainly supervised learning Language modeling
More informationEE562 ARTIFICIAL INTELLIGENCE FOR ENGINEERS
EE562 ARTIFICIAL INTELLIGENCE FOR ENGINEERS Lecture 16, 6/1/2005 University of Washington, Department of Electrical Engineering Spring 2005 Instructor: Professor Jeff A. Bilmes Uncertainty & Bayesian Networks
More informationDirected and Undirected Graphical Models
Directed and Undirected Graphical Models Adrian Weller MLSALT4 Lecture Feb 26, 2016 With thanks to David Sontag (NYU) and Tony Jebara (Columbia) for use of many slides and illustrations For more information,
More informationNaïve Bayes classification
Naïve Bayes classification 1 Probability theory Random variable: a variable whose possible values are numerical outcomes of a random phenomenon. Examples: A person s height, the outcome of a coin toss
More informationBayesian Networks Inference with Probabilistic Graphical Models
4190.408 2016-Spring Bayesian Networks Inference with Probabilistic Graphical Models Byoung-Tak Zhang intelligence Lab Seoul National University 4190.408 Artificial (2016-Spring) 1 Machine Learning? Learning
More informationDirected Graphical Models or Bayesian Networks
Directed Graphical Models or Bayesian Networks Le Song Machine Learning II: Advanced Topics CSE 8803ML, Spring 2012 Bayesian Networks One of the most exciting recent advancements in statistical AI Compact
More informationConditional Independence and Factorization
Conditional Independence and Factorization Seungjin Choi Department of Computer Science and Engineering Pohang University of Science and Technology 77 Cheongam-ro, Nam-gu, Pohang 37673, Korea seungjin@postech.ac.kr
More informationProbability. Machine Learning and Pattern Recognition. Chris Williams. School of Informatics, University of Edinburgh. August 2014
Probability Machine Learning and Pattern Recognition Chris Williams School of Informatics, University of Edinburgh August 2014 (All of the slides in this course have been adapted from previous versions
More informationProbabilistic modeling. The slides are closely adapted from Subhransu Maji s slides
Probabilistic modeling The slides are closely adapted from Subhransu Maji s slides Overview So far the models and algorithms you have learned about are relatively disconnected Probabilistic modeling framework
More informationBased on slides by Richard Zemel
CSC 412/2506 Winter 2018 Probabilistic Learning and Reasoning Lecture 3: Directed Graphical Models and Latent Variables Based on slides by Richard Zemel Learning outcomes What aspects of a model can we
More informationUndirected Graphical Models
Outline Hong Chang Institute of Computing Technology, Chinese Academy of Sciences Machine Learning Methods (Fall 2012) Outline Outline I 1 Introduction 2 Properties Properties 3 Generative vs. Conditional
More informationBayesian Networks. Motivation
Bayesian Networks Computer Sciences 760 Spring 2014 http://pages.cs.wisc.edu/~dpage/cs760/ Motivation Assume we have five Boolean variables,,,, The joint probability is,,,, How many state configurations
More informationBayesian Learning. HT2015: SC4 Statistical Data Mining and Machine Learning. Maximum Likelihood Principle. The Bayesian Learning Framework
HT5: SC4 Statistical Data Mining and Machine Learning Dino Sejdinovic Department of Statistics Oxford http://www.stats.ox.ac.uk/~sejdinov/sdmml.html Maximum Likelihood Principle A generative model for
More informationIntroduction to Artificial Intelligence. Unit # 11
Introduction to Artificial Intelligence Unit # 11 1 Course Outline Overview of Artificial Intelligence State Space Representation Search Techniques Machine Learning Logic Probabilistic Reasoning/Bayesian
More informationProbabilistic Models. Models describe how (a portion of) the world works
Probabilistic Models Models describe how (a portion of) the world works Models are always simplifications May not account for every variable May not account for all interactions between variables All models
More informationCS839: Probabilistic Graphical Models. Lecture 7: Learning Fully Observed BNs. Theo Rekatsinas
CS839: Probabilistic Graphical Models Lecture 7: Learning Fully Observed BNs Theo Rekatsinas 1 Exponential family: a basic building block For a numeric random variable X p(x ) =h(x)exp T T (x) A( ) = 1
More informationSTA 4273H: Statistical Machine Learning
STA 4273H: Statistical Machine Learning Russ Salakhutdinov Department of Computer Science! Department of Statistical Sciences! rsalakhu@cs.toronto.edu! h0p://www.cs.utoronto.ca/~rsalakhu/ Lecture 7 Approximate
More informationNPFL108 Bayesian inference. Introduction. Filip Jurčíček. Institute of Formal and Applied Linguistics Charles University in Prague Czech Republic
NPFL108 Bayesian inference Introduction Filip Jurčíček Institute of Formal and Applied Linguistics Charles University in Prague Czech Republic Home page: http://ufal.mff.cuni.cz/~jurcicek Version: 21/02/2014
More information9/12/17. Types of learning. Modeling data. Supervised learning: Classification. Supervised learning: Regression. Unsupervised learning: Clustering
Types of learning Modeling data Supervised: we know input and targets Goal is to learn a model that, given input data, accurately predicts target data Unsupervised: we know the input only and want to make
More informationProbability and Information Theory. Sargur N. Srihari
Probability and Information Theory Sargur N. srihari@cedar.buffalo.edu 1 Topics in Probability and Information Theory Overview 1. Why Probability? 2. Random Variables 3. Probability Distributions 4. Marginal
More informationBayesian Methods in Artificial Intelligence
WDS'10 Proceedings of Contributed Papers, Part I, 25 30, 2010. ISBN 978-80-7378-139-2 MATFYZPRESS Bayesian Methods in Artificial Intelligence M. Kukačka Charles University, Faculty of Mathematics and Physics,
More informationOutline. CSE 573: Artificial Intelligence Autumn Bayes Nets: Big Picture. Bayes Net Semantics. Hidden Markov Models. Example Bayes Net: Car
CSE 573: Artificial Intelligence Autumn 2012 Bayesian Networks Dan Weld Many slides adapted from Dan Klein, Stuart Russell, Andrew Moore & Luke Zettlemoyer Outline Probabilistic models (and inference)
More informationA Tutorial on Learning with Bayesian Networks
A utorial on Learning with Bayesian Networks David Heckerman Presented by: Krishna V Chengavalli April 21 2003 Outline Introduction Different Approaches Bayesian Networks Learning Probabilities and Structure
More informationDirected Probabilistic Graphical Models CMSC 678 UMBC
Directed Probabilistic Graphical Models CMSC 678 UMBC Announcement 1: Assignment 3 Due Wednesday April 11 th, 11:59 AM Any questions? Announcement 2: Progress Report on Project Due Monday April 16 th,
More informationStatistical Approaches to Learning and Discovery
Statistical Approaches to Learning and Discovery Graphical Models Zoubin Ghahramani & Teddy Seidenfeld zoubin@cs.cmu.edu & teddy@stat.cmu.edu CALD / CS / Statistics / Philosophy Carnegie Mellon University
More informationProbabilistic Reasoning. (Mostly using Bayesian Networks)
Probabilistic Reasoning (Mostly using Bayesian Networks) Introduction: Why probabilistic reasoning? The world is not deterministic. (Usually because information is limited.) Ways of coping with uncertainty
More informationIntroduction: MLE, MAP, Bayesian reasoning (28/8/13)
STA561: Probabilistic machine learning Introduction: MLE, MAP, Bayesian reasoning (28/8/13) Lecturer: Barbara Engelhardt Scribes: K. Ulrich, J. Subramanian, N. Raval, J. O Hollaren 1 Classifiers In this
More informationDirected and Undirected Graphical Models
Directed and Undirected Davide Bacciu Dipartimento di Informatica Università di Pisa bacciu@di.unipi.it Machine Learning: Neural Networks and Advanced Models (AA2) Last Lecture Refresher Lecture Plan Directed
More informationLecture : Probabilistic Machine Learning
Lecture : Probabilistic Machine Learning Riashat Islam Reasoning and Learning Lab McGill University September 11, 2018 ML : Many Methods with Many Links Modelling Views of Machine Learning Machine Learning
More informationUndirected Graphical Models: Markov Random Fields
Undirected Graphical Models: Markov Random Fields 40-956 Advanced Topics in AI: Probabilistic Graphical Models Sharif University of Technology Soleymani Spring 2015 Markov Random Field Structure: undirected
More informationSTA 4273H: Statistical Machine Learning
STA 4273H: Statistical Machine Learning Russ Salakhutdinov Department of Statistics! rsalakhu@utstat.toronto.edu! http://www.utstat.utoronto.ca/~rsalakhu/ Sidney Smith Hall, Room 6002 Lecture 11 Project
More informationOutline. CSE 573: Artificial Intelligence Autumn Agent. Partial Observability. Markov Decision Process (MDP) 10/31/2012
CSE 573: Artificial Intelligence Autumn 2012 Reasoning about Uncertainty & Hidden Markov Models Daniel Weld Many slides adapted from Dan Klein, Stuart Russell, Andrew Moore & Luke Zettlemoyer 1 Outline
More informationBayesian Methods for Machine Learning
Bayesian Methods for Machine Learning CS 584: Big Data Analytics Material adapted from Radford Neal s tutorial (http://ftp.cs.utoronto.ca/pub/radford/bayes-tut.pdf), Zoubin Ghahramni (http://hunch.net/~coms-4771/zoubin_ghahramani_bayesian_learning.pdf),
More informationIntroduction to Probabilistic Machine Learning
Introduction to Probabilistic Machine Learning Piyush Rai Dept. of CSE, IIT Kanpur (Mini-course 1) Nov 03, 2015 Piyush Rai (IIT Kanpur) Introduction to Probabilistic Machine Learning 1 Machine Learning
More informationCS 188: Artificial Intelligence Spring Announcements
CS 188: Artificial Intelligence Spring 2011 Lecture 16: Bayes Nets IV Inference 3/28/2011 Pieter Abbeel UC Berkeley Many slides over this course adapted from Dan Klein, Stuart Russell, Andrew Moore Announcements
More informationCSE 473: Artificial Intelligence Autumn Topics
CSE 473: Artificial Intelligence Autumn 2014 Bayesian Networks Learning II Dan Weld Slides adapted from Jack Breese, Dan Klein, Daphne Koller, Stuart Russell, Andrew Moore & Luke Zettlemoyer 1 473 Topics
More informationBayes Networks. CS540 Bryan R Gibson University of Wisconsin-Madison. Slides adapted from those used by Prof. Jerry Zhu, CS540-1
Bayes Networks CS540 Bryan R Gibson University of Wisconsin-Madison Slides adapted from those used by Prof. Jerry Zhu, CS540-1 1 / 59 Outline Joint Probability: great for inference, terrible to obtain
More informationGraphical Models. Andrea Passerini Statistical relational learning. Graphical Models
Andrea Passerini passerini@disi.unitn.it Statistical relational learning Probability distributions Bernoulli distribution Two possible values (outcomes): 1 (success), 0 (failure). Parameters: p probability
More informationNaïve Bayes classification. p ij 11/15/16. Probability theory. Probability theory. Probability theory. X P (X = x i )=1 i. Marginal Probability
Probability theory Naïve Bayes classification Random variable: a variable whose possible values are numerical outcomes of a random phenomenon. s: A person s height, the outcome of a coin toss Distinguish
More informationECE 6504: Advanced Topics in Machine Learning Probabilistic Graphical Models and Large-Scale Learning
ECE 6504: Advanced Topics in Machine Learning Probabilistic Graphical Models and Large-Scale Learning Topics Summary of Class Advanced Topics Dhruv Batra Virginia Tech HW1 Grades Mean: 28.5/38 ~= 74.9%
More informationProbabilistic Graphical Models
Probabilistic Graphical Models Brown University CSCI 295-P, Spring 213 Prof. Erik Sudderth Lecture 11: Inference & Learning Overview, Gaussian Graphical Models Some figures courtesy Michael Jordan s draft
More informationGraphical Models 359
8 Graphical Models Probabilities play a central role in modern pattern recognition. We have seen in Chapter 1 that probability theory can be expressed in terms of two simple equations corresponding to
More informationProbabilistic Models
Bayes Nets 1 Probabilistic Models Models describe how (a portion of) the world works Models are always simplifications May not account for every variable May not account for all interactions between variables
More informationProbabilistic Graphical Models and Bayesian Networks. Artificial Intelligence Bert Huang Virginia Tech
Probabilistic Graphical Models and Bayesian Networks Artificial Intelligence Bert Huang Virginia Tech Concept Map for Segment Probabilistic Graphical Models Probabilistic Time Series Models Particle Filters
More informationCovariance. if X, Y are independent
Review: probability Monty Hall, weighted dice Frequentist v. Bayesian Independence Expectations, conditional expectations Exp. & independence; linearity of exp. Estimator (RV computed from sample) law
More informationProbabilistic Graphical Models: MRFs and CRFs. CSE628: Natural Language Processing Guest Lecturer: Veselin Stoyanov
Probabilistic Graphical Models: MRFs and CRFs CSE628: Natural Language Processing Guest Lecturer: Veselin Stoyanov Why PGMs? PGMs can model joint probabilities of many events. many techniques commonly
More informationLearning Bayesian network : Given structure and completely observed data
Learning Bayesian network : Given structure and completely observed data Probabilistic Graphical Models Sharif University of Technology Spring 2017 Soleymani Learning problem Target: true distribution
More informationIntroduction to Probability and Statistics (Continued)
Introduction to Probability and Statistics (Continued) Prof. icholas Zabaras Center for Informatics and Computational Science https://cics.nd.edu/ University of otre Dame otre Dame, Indiana, USA Email:
More informationPROBABILITY DISTRIBUTIONS. J. Elder CSE 6390/PSYC 6225 Computational Modeling of Visual Perception
PROBABILITY DISTRIBUTIONS Credits 2 These slides were sourced and/or modified from: Christopher Bishop, Microsoft UK Parametric Distributions 3 Basic building blocks: Need to determine given Representation:
More informationBayes Nets: Independence
Bayes Nets: Independence [These slides were created by Dan Klein and Pieter Abbeel for CS188 Intro to AI at UC Berkeley. All CS188 materials are available at http://ai.berkeley.edu.] Bayes Nets A Bayes
More informationProbabilistic Graphical Models
Probabilistic Graphical Models Introduction. Basic Probability and Bayes Volkan Cevher, Matthias Seeger Ecole Polytechnique Fédérale de Lausanne 26/9/2011 (EPFL) Graphical Models 26/9/2011 1 / 28 Outline
More informationT Machine Learning: Basic Principles
Machine Learning: Basic Principles Bayesian Networks Laboratory of Computer and Information Science (CIS) Department of Computer Science and Engineering Helsinki University of Technology (TKK) Autumn 2007
More informationMachine Learning using Bayesian Approaches
Machine Learning using Bayesian Approaches Sargur N. Srihari University at Buffalo, State University of New York 1 Outline 1. Progress in ML and PR 2. Fully Bayesian Approach 1. Probability theory Bayes
More informationMachine Learning. Probability Basics. Marc Toussaint University of Stuttgart Summer 2014
Machine Learning Probability Basics Basic definitions: Random variables, joint, conditional, marginal distribution, Bayes theorem & examples; Probability distributions: Binomial, Beta, Multinomial, Dirichlet,
More informationRepresentation. Stefano Ermon, Aditya Grover. Stanford University. Lecture 2
Representation Stefano Ermon, Aditya Grover Stanford University Lecture 2 Stefano Ermon, Aditya Grover (AI Lab) Deep Generative Models Lecture 2 1 / 32 Learning a generative model We are given a training
More informationLecture 8: Bayesian Networks
Lecture 8: Bayesian Networks Bayesian Networks Inference in Bayesian Networks COMP-652 and ECSE 608, Lecture 8 - January 31, 2017 1 Bayes nets P(E) E=1 E=0 0.005 0.995 E B P(B) B=1 B=0 0.01 0.99 E=0 E=1
More information