10601 Machine Learning
|
|
- Poppy George
- 5 years ago
- Views:
Transcription
1 10601 Machine Learning September 2, 2009 Recitation 2 Öznur Taştan 1
2 Logistics Homework 2 is going to be out tomorrow. It is due on Sep 16, Wed. There is no class on Monday Sep 7 th (Labor day) Those who have not return Homework 1 yet For details of how to submit the homework policy please check :
3 Outline We will review Some probability and statistics Some graphical models We will not go over Homework 1 Since the grace period has not ended yet. Solutions will be up next week on the web page.
4 We ll play a game: Catch the goof! I ll be the sloppy TA will make intentional mistakes Slides with mistakes are marked with You ll catch those mistakes and correct me! Correct slides are marked with
5 Catch the goof!!
6 Law of total probability Given two discrete random variables X and Y X takes values in { x x } 1,, m y1,, yn Y takes values in{ } PY ( = y) = P( X= x, Y= y) j i j j PY ( = y) = P( X= x Y= y) PY ( = y) j i j j j
7 Law of total probability Given two discrete random variables X and Y X takes values in { x x } 1,, m y1,, yn Y takes values in{ } PY ( = y) = P( X= x, Y= y) j i j j PY ( = y) = P( X= x Y= y) PY ( = y) j i j j j
8 Law of total probability Given two discrete random variables X and Y X takes values in { x x } 1,, m y1,, yn Y takes values in{ } P( X = x i ) PY ( = y) = P( X= x, Y= y) j i j j PY ( = y) = P( X= x Y= y) PY ( = y) P( X = x i ) j i j j j
9 Law of total probability Given two discrete random variables X and Y X takes values in { x x } 1,, m y1,, yn Y takes values in{ } PX ( = x) = PX ( = xy, = y) i i j j PX ( = x) = PX ( = x Y= y) PY ( = y) i i j j j
10 Law of total probability Given two discrete random variables X and Y PX ( = x) = PX ( = xy, = y) Marginal probability i i j j Joint probability = PX ( = xi Y= yj) PY ( = yj) j Conditional probability of X conditioned on Y
11 Law of total probability Given two discrete random variables X and Y Formulas are fine. Anything wrong with the names? PX ( = x) = PX ( = xy, = y) Marginal probability i i j j Joint probability = PX ( = xi Y= yj) PY ( = yj) j Conditional probability of X conditioned on Y
12 Law of total probability Given two discrete random variables X and Y PX ( = x) = PX ( = xy, = y) i i j j Joint probability of X,Y Marginal probability = PX ( = xi Y= yj) PY ( = yj) j Conditional probability of X conditioned on Y Marginal probability
13 In a strange world Two discrete random variables X and Y take binary values Joint probabilities P( X = 0, Y = 1) = 0.2 P( X = 0, Y = 0) = 0.2 P( X = 1, Y = 0) = 0.5 P( X = 1, Y = 1) = 0.5
14 In a strange world Two discrete random variables X and Y take binary values Joint probabilities P( X = 0, Y = 1) = 0.2 Should sum up to 1 P( X = 0, Y = 0) = 0.2 P( X = 1, Y = 0) = 0.5 P( X = 1, Y = 1) = 0.5
15 The world seems fine Two discrete random variables X and Y take binary values Joint probabilities P( X = 0, Y = 1) = 0.2 P( X = 0, Y = 0) = 0.2 P( X = 1, Y = 0) = 0.3 P( X = 1, Y = 1) = 0.3
16 What about the marginals? Joint probabilities P( X = 0, Y = 1) = 0.2 P( X = 0, Y = 0) = 0.2 P( X = 1, Y = 0) = 0.3 P( X = 1, Y = 1) = 0.3 Marginal probabilities P( X = 0) = 0.2 PX ( = 1) = 0.8 P( Y = 0) = 0.5 PY ( = 1) = 0.5
17 This is a strange world Joint probabilities P( X = 0, Y = 1) = 0.2 P( X = 0, Y = 0) = 0.2 P( X = 1, Y = 0) = 0.3 P( X = 1, Y = 1) = 0.3 Marginal probabilities P( X = 0) = 0.2 PX ( = 1) = 0.8 P( Y = 0) = 0.5 PY ( = 1) = 0.5
18 In a strange world Joint probabilities P( X = 0, Y = 1) = 0.2 P( X = 0, Y = 0) = 0.2 P( X = 1, Y = 0) = 0.3 P( X = 1, Y = 1) = 0.3 Marginal probabilities P( X = 0) = 0.2 PX ( = 1) = 0.8 P( Y = 0) = 0.5 PY ( = 1) = 0.5
19 This is a strange world Joint probabilities P( X = 0, Y = 1) = 0.2 P( X = 0, Y = 0) = 0.2 P( X = 1, Y = 0) = 0.3 P( X = 1, Y = 1) = 0.3 Marginal probabilities P( X = 0) = 0.2 PX ( = 1) = 0.8 P( Y = 0) = 0.5 PY ( = 1) = 0.5
20 Let s have a simple problem Joint probabilities P( X = 0, Y = 1) = 0.2 P( X = 0, Y = 0) = 0.2 P( X = 1, Y = 0) = 0.3 P( X = 1, Y = 1) = 0.3 Marginal probabilities P( X = 0) = 0.4 PX ( = 1) = 0.6 P( Y = 0) = 0.5 PY ( = 1) = 0.5 P( X = 0) = P( X = 0, Y = 0) + P( X = 1, Y = 1) =0.4
21 Conditional probabilities What is the complementary event of P(X=0 Y=1)? P(X=1 Y=1) OR P(X=0 Y=0)
22 Conditional probabilities What is the complementary event of P(X=0 Y=1)? P(X=1 Y=1) OR P(X=0 Y=0)
23 The game ends here.
24 Independent number of parameters Assume X and Y take Boolean values {0,1}: How many independent parameters do you need to fully specify: marginal probability of X? the joint probability of P(X,Y)? the conditional probability of P(X Y)?
25 Independent number of parameters Assume X and Y take Boolean values {0,1}: How many independent parameters do you need to fully specify: marginal probability of X? P(X=0) 1 parameter only [ because P(X=1)+P(X=0)=1 ] the joint probability of P(X,Y)? P(X=0, Y=0) 3 parameters P(X=0, Y=1) P(X=1, Y=0) the conditional probability of P(X Y)?
26 Number of parameters Assume X and Y take Boolean values {0,1}? How many independent parameters do you need to fully specify marginal probability of X? P(X=0) 1 parameter only P(X=1)= 1 P(X=0) How many independent parameters do you need to fully specify the joint probability of P(X,Y)? P(X=0, Y=0) 3 parameters P(X=0, Y=1) P(X=1, Y=0) How many independent parameters do you need to fully specify the conditional probability of P(X Y)? P(X=0 Y=0) 2 parameters P(X=0 Y=1)
27 Number of parameters What about P(X Y,Z), how many independent parameters do you need to be able to fully specify the probabilities? Assume each RV takes: P(X Y,Z) m values n values q values
28 Number of parameters What about P(X Y,Z), how many independent parameters do you need to be able to fully specify the probabilities? Assume each RV takes: P(X Y,Z) m values n values q values Number of independent parameters: (m 1)*nq
29 Graphical models A graphical model is a way of representing probabilistic relationships between random variables Variables are represented by nodes: Edges indicates probabilistic relationships: P(X,X..X ) = P(X Pa(X )) 1 2 n i i i= 1 You miss the bus Arrive class late
30 Serial connection Is X and Z independent? X X Z? Y Z
31 Serial connection Is X and Z independent? X X Z Y X and Z are not independent Z
32 Serial connection Is X conditionally independent of Z given Y? X X Z Y? Y Z
33 Serial connection Is X conditionally independent of Z given Y? X X Z Y Y Yes they are independent Z
34 How can we show it? Is X conditionally independent of Z given Y? X P( XYZ,, ) = PXPY ( ) ( XPZ ) ( Y) Y Z P( Z X, Y) PXYZ (,, ) = PXY (, ) P( XPY ) ( XPZY ) ( ) = PXPY ( ) ( X) = PZ ( Y) X Z Y
35 An example case Studied late last night Wake up late Arrive class late
36 Common cause Age Z X Y Shoe Size Gray Hair X and Y are not marginally independent X and Y are conditionally independent given Z
37 Explaining away Flu X Z Allergy Y Sneeze X and Z marginally independent X and Z conditionally dependent given Y
38 D separation X and Z are conditionally independent given Y if Y d separates X and Z X X X X Y Y Y Y Neither Y nor its descendants should be observed Z Z Z Z Path between X and Z is blocked by Y
39 D separation example Is B, C independent given A?
40 D separation example Is B, C independent given A? Yes
41 D separation example Observed, A blocks the path Is B, C independent given A? Yes
42 Observed, A blocks the path Is B, C independent given A? Yes not observed neither its descendants
43 D separation example Is A, F independent given E?
44 Is A, F independent given E? Yes
45 Is A, F independent given E? Yes
46 Is C, D independent given F?
47 Is C, D independent given F? No
48 Is A, G independent given B and F?
49 Is A, G independent given B and F? Yes
50 Naïve Bayes Model J D C R J: The person is a junior D: The person knows calculus C: The person leaves in campus R: Saw the Return of the King more than once
51 Naïve Bayes Model J What parameters are stored? D C R J: The person is a junior D: The person knows calculus C: The person leaves in campus R: Saw the Return of the King more than once
52 Naïve Bayes Model J P(J)= D C R P(D/J=1)= P(D/J=0)= P(C/J=1)= P(C/J=0)= P(R/J=1)= P(R/J=0)= J : The person is a junior D: The person knows calculus C: The person leaves in campus R: Saw the Return of the King more than once
53 Naïve Bayes Model J P(J)= D C R P(D/J=1)= P(D/J=0)= P(C/J=1)= P(C/J=0)= P(R/J=1)= P(R/J=0)= J: The person is a junior D: The person knows calculus C: The person leaves in campus R: Saw the Return of the King more than once
54 We have the structure how do we get the CPTs? Estimate them from observed data Are you a junior? Do you know calculus? Do you live in campus? Have you seen 'Return of the King more than once? Student Student Student Student Student Student Student Student Student Student Student Student Student Student Student Student Student Student Student Student
55 What is the probability that he is a Junior? Naïve Bayes Model P(J)= J: The person is a junior D: The person knows calculus C: The person leaves in campus R: Saw the Return of the King more than once J D C R P(R/J)= P(R/~J)= P(C/J)= P(C/J)= Suppose a new person come and says: P(C/~J)= I don t know calculus P(C/~J)= I live in campus I have seen The return of the king five times
56 Naïve Bayes Model J Suppose a person says: I don t know calculus D=0 I live in campus C=1 I have not seen The return of the king five times R=1 D C R What is the probability that he is a Junior? P(J=1/D=0,C=1,R=1)
57 What is the probability that he is a Junior? P(J=1/D=0,C=1,R=1) = P(J=1,D=0,C=1,R=1) P(D = 0,C = 1,R = 1) = P(J = 1)P(C= 1/J= 1)P(R = 1/J= 1)P(D= 0/J= 1) P(D = 0,C = 1,R = 1) To calculate this marginalize over J J D C R
58 Naïve Bayes Model P(J=1/D=0,C=1,R=1) P(J = 1)P(C= 1/J= 1)P(R = 1/J= 1)P(D= 0/J= 1) = P(D = 0,C = 1,R = 1)
Introduction to probability and statistics
Introduction to probability and statistics Alireza Fotuhi Siahpirani & Brittany Baur sroy@biostat.wisc.edu Computational Network Biology Biostatistics & Medical Informatics 826 https://compnetbiocourse.discovery.wisc.edu
More information1 Expectation Maximization
Introduction Expectation-Maximization Bibliographical notes 1 Expectation Maximization Daniel Khashabi 1 khashab2@illinois.edu 1.1 Introduction Consider the problem of parameter learning by maximizing
More informationLecture 8. Probabilistic Reasoning CS 486/686 May 25, 2006
Lecture 8 Probabilistic Reasoning CS 486/686 May 25, 2006 Outline Review probabilistic inference, independence and conditional independence Bayesian networks What are they What do they mean How do we create
More informationConditional Independence and Factorization
Conditional Independence and Factorization Seungjin Choi Department of Computer Science and Engineering Pohang University of Science and Technology 77 Cheongam-ro, Nam-gu, Pohang 37673, Korea seungjin@postech.ac.kr
More informationCS 188: Artificial Intelligence. Bayes Nets
CS 188: Artificial Intelligence Probabilistic Inference: Enumeration, Variable Elimination, Sampling Pieter Abbeel UC Berkeley Many slides over this course adapted from Dan Klein, Stuart Russell, Andrew
More informationRecall from last time: Conditional probabilities. Lecture 2: Belief (Bayesian) networks. Bayes ball. Example (continued) Example: Inference problem
Recall from last time: Conditional probabilities Our probabilistic models will compute and manipulate conditional probabilities. Given two random variables X, Y, we denote by Lecture 2: Belief (Bayesian)
More informationDirected Graphical Models or Bayesian Networks
Directed Graphical Models or Bayesian Networks Le Song Machine Learning II: Advanced Topics CSE 8803ML, Spring 2012 Bayesian Networks One of the most exciting recent advancements in statistical AI Compact
More informationProbabilistic Graphical Models and Bayesian Networks. Artificial Intelligence Bert Huang Virginia Tech
Probabilistic Graphical Models and Bayesian Networks Artificial Intelligence Bert Huang Virginia Tech Concept Map for Segment Probabilistic Graphical Models Probabilistic Time Series Models Particle Filters
More informationCOS402- Artificial Intelligence Fall Lecture 10: Bayesian Networks & Exact Inference
COS402- Artificial Intelligence Fall 2015 Lecture 10: Bayesian Networks & Exact Inference Outline Logical inference and probabilistic inference Independence and conditional independence Bayes Nets Semantics
More informationBayes Nets: Independence
Bayes Nets: Independence [These slides were created by Dan Klein and Pieter Abbeel for CS188 Intro to AI at UC Berkeley. All CS188 materials are available at http://ai.berkeley.edu.] Bayes Nets A Bayes
More information1. what conditional independencies are implied by the graph. 2. whether these independecies correspond to the probability distribution
NETWORK ANALYSIS Lourens Waldorp PROBABILITY AND GRAPHS The objective is to obtain a correspondence between the intuitive pictures (graphs) of variables of interest and the probability distributions of
More informationLogistic Regression Review Fall 2012 Recitation. September 25, 2012 TA: Selen Uguroglu
Logistic Regression Review 10-601 Fall 2012 Recitation September 25, 2012 TA: Selen Uguroglu!1 Outline Decision Theory Logistic regression Goal Loss function Inference Gradient Descent!2 Training Data
More informationProbabilistic Graphical Networks: Definitions and Basic Results
This document gives a cursory overview of Probabilistic Graphical Networks. The material has been gleaned from different sources. I make no claim to original authorship of this material. Bayesian Graphical
More informationLearning P-maps Param. Learning
Readings: K&F: 3.3, 3.4, 16.1, 16.2, 16.3, 16.4 Learning P-maps Param. Learning Graphical Models 10708 Carlos Guestrin Carnegie Mellon University September 24 th, 2008 10-708 Carlos Guestrin 2006-2008
More informationBayesian Networks Representation
Bayesian Networks Representation Machine Learning 10701/15781 Carlos Guestrin Carnegie Mellon University March 19 th, 2007 Handwriting recognition Character recognition, e.g., kernel SVMs a c z rr r r
More informationSome Probability and Statistics
Some Probability and Statistics David M. Blei COS424 Princeton University February 13, 2012 Card problem There are three cards Red/Red Red/Black Black/Black I go through the following process. Close my
More informationTópicos Especiais em Modelagem e Análise - Aprendizado por Máquina CPS863
Tópicos Especiais em Modelagem e Análise - Aprendizado por Máquina CPS863 Daniel, Edmundo, Rosa Terceiro trimestre de 2012 UFRJ - COPPE Programa de Engenharia de Sistemas e Computação Bayesian Networks
More informationGraphical Models - Part I
Graphical Models - Part I Oliver Schulte - CMPT 726 Bishop PRML Ch. 8, some slides from Russell and Norvig AIMA2e Outline Probabilistic Models Bayesian Networks Markov Random Fields Inference Outline Probabilistic
More informationToday. Why do we need Bayesian networks (Bayes nets)? Factoring the joint probability. Conditional independence
Bayesian Networks Today Why do we need Bayesian networks (Bayes nets)? Factoring the joint probability Conditional independence What is a Bayes net? Mostly, we discuss discrete r.v. s What can you do with
More informationCSEP 573: Artificial Intelligence
CSEP 573: Artificial Intelligence Bayesian Networks: Inference Ali Farhadi Many slides over the course adapted from either Luke Zettlemoyer, Pieter Abbeel, Dan Klein, Stuart Russell or Andrew Moore 1 Outline
More informationBayesian networks (1) Lirong Xia
Bayesian networks (1) Lirong Xia Random variables and joint distributions Ø A random variable is a variable with a domain Random variables: capital letters, e.g. W, D, L values: small letters, e.g. w,
More informationCMPSCI 240: Reasoning about Uncertainty
CMPSCI 240: Reasoning about Uncertainty Lecture 17: Representing Joint PMFs and Bayesian Networks Andrew McGregor University of Massachusetts Last Compiled: April 7, 2017 Warm Up: Joint distributions Recall
More informationDirected Graphical Models
CS 2750: Machine Learning Directed Graphical Models Prof. Adriana Kovashka University of Pittsburgh March 28, 2017 Graphical Models If no assumption of independence is made, must estimate an exponential
More informationMathematical Foundations
Mathematical Foundations Introduction to Data Science Algorithms Michael Paul and Jordan Boyd-Graber JANUARY 23, 2017 Introduction to Data Science Algorithms Boyd-Graber and Paul Mathematical Foundations
More informationRepresentation. Stefano Ermon, Aditya Grover. Stanford University. Lecture 2
Representation Stefano Ermon, Aditya Grover Stanford University Lecture 2 Stefano Ermon, Aditya Grover (AI Lab) Deep Generative Models Lecture 2 1 / 32 Learning a generative model We are given a training
More informationFusion in simple models
Fusion in simple models Peter Antal antal@mit.bme.hu A.I. February 8, 2018 1 Basic concepts of probability theory Joint distribution Conditional probability Bayes rule Chain rule Marginalization General
More informationIntroduction to Bayesian Learning
Course Information Introduction Introduction to Bayesian Learning Davide Bacciu Dipartimento di Informatica Università di Pisa bacciu@di.unipi.it Apprendimento Automatico: Fondamenti - A.A. 2016/2017 Outline
More informationNaïve Bayes Introduction to Machine Learning. Matt Gormley Lecture 18 Oct. 31, 2018
10-601 Introduction to Machine Learning Machine Learning Department School of Computer Science Carnegie Mellon University Naïve Bayes Matt Gormley Lecture 18 Oct. 31, 2018 1 Reminders Homework 6: PAC Learning
More informationProbabilistic Graphical Models
Probabilistic Graphical Models Lecture 4 Learning Bayesian Networks CS/CNS/EE 155 Andreas Krause Announcements Another TA: Hongchao Zhou Please fill out the questionnaire about recitations Homework 1 out.
More informationSDS 321: Introduction to Probability and Statistics
SDS 321: Introduction to Probability and Statistics Lecture 17: Continuous random variables: conditional PDF Purnamrita Sarkar Department of Statistics and Data Science The University of Texas at Austin
More informationFinal Exam. December 11 th, This exam booklet contains five problems, out of which you are expected to answer four problems of your choice.
CS446: Machine Learning Fall 2012 Final Exam December 11 th, 2012 This is a closed book exam. Everything you need in order to solve the problems is supplied in the body of this exam. Note that there is
More informationUVA CS / Introduc8on to Machine Learning and Data Mining
UVA CS 4501-001 / 6501 007 Introduc8on to Machine Learning and Data Mining Lecture 13: Probability and Sta3s3cs Review (cont.) + Naïve Bayes Classifier Yanjun Qi / Jane, PhD University of Virginia Department
More informationDirected Graphical Models
Directed Graphical Models Instructor: Alan Ritter Many Slides from Tom Mitchell Graphical Models Key Idea: Conditional independence assumptions useful but Naïve Bayes is extreme! Graphical models express
More informationProbabilistic Machine Learning
Probabilistic Machine Learning Bayesian Nets, MCMC, and more Marek Petrik 4/18/2017 Based on: P. Murphy, K. (2012). Machine Learning: A Probabilistic Perspective. Chapter 10. Conditional Independence Independent
More informationProbabilistic Models. Models describe how (a portion of) the world works
Probabilistic Models Models describe how (a portion of) the world works Models are always simplifications May not account for every variable May not account for all interactions between variables All models
More informationBayes Nets III: Inference
1 Hal Daumé III (me@hal3.name) Bayes Nets III: Inference Hal Daumé III Computer Science University of Maryland me@hal3.name CS 421: Introduction to Artificial Intelligence 10 Apr 2012 Many slides courtesy
More informationMachine Learning (CS 567) Lecture 2
Machine Learning (CS 567) Lecture 2 Time: T-Th 5:00pm - 6:20pm Location: GFS118 Instructor: Sofus A. Macskassy (macskass@usc.edu) Office: SAL 216 Office hours: by appointment Teaching assistant: Cheol
More informationAn Introduction to Bayesian Machine Learning
1 An Introduction to Bayesian Machine Learning José Miguel Hernández-Lobato Department of Engineering, Cambridge University April 8, 2013 2 What is Machine Learning? The design of computational systems
More informationIntroduction to Bayesian Learning. Machine Learning Fall 2018
Introduction to Bayesian Learning Machine Learning Fall 2018 1 What we have seen so far What does it mean to learn? Mistake-driven learning Learning by counting (and bounding) number of mistakes PAC learnability
More informationRapid Introduction to Machine Learning/ Deep Learning
Rapid Introduction to Machine Learning/ Deep Learning Hyeong In Choi Seoul National University 1/32 Lecture 5a Bayesian network April 14, 2016 2/32 Table of contents 1 1. Objectives of Lecture 5a 2 2.Bayesian
More informationConditional probabilities and graphical models
Conditional probabilities and graphical models Thomas Mailund Bioinformatics Research Centre (BiRC), Aarhus University Probability theory allows us to describe uncertainty in the processes we model within
More informationReview of probability
Review of probability Computer Sciences 760 Spring 2014 http://pages.cs.wisc.edu/~dpage/cs760/ Goals for the lecture you should understand the following concepts definition of probability random variables
More informationNaïve Bayes classification
Naïve Bayes classification 1 Probability theory Random variable: a variable whose possible values are numerical outcomes of a random phenomenon. Examples: A person s height, the outcome of a coin toss
More informationIntroduction to Bayes Nets. CS 486/686: Introduction to Artificial Intelligence Fall 2013
Introduction to Bayes Nets CS 486/686: Introduction to Artificial Intelligence Fall 2013 1 Introduction Review probabilistic inference, independence and conditional independence Bayesian Networks - - What
More informationCPSC 540: Machine Learning
CPSC 540: Machine Learning Undirected Graphical Models Mark Schmidt University of British Columbia Winter 2016 Admin Assignment 3: 2 late days to hand it in today, Thursday is final day. Assignment 4:
More informationCS 188: Artificial Intelligence Spring Announcements
CS 188: Artificial Intelligence Spring 2011 Lecture 14: Bayes Nets II Independence 3/9/2011 Pieter Abbeel UC Berkeley Many slides over the course adapted from Dan Klein, Stuart Russell, Andrew Moore Announcements
More informationMachine Learning Lecture 14
Many slides adapted from B. Schiele, S. Roth, Z. Gharahmani Machine Learning Lecture 14 Undirected Graphical Models & Inference 23.06.2015 Bastian Leibe RWTH Aachen http://www.vision.rwth-aachen.de/ leibe@vision.rwth-aachen.de
More informationBayesian Networks (Part II)
10-601 Introduction to Machine Learning Machine Learning Department School of Computer Science Carnegie Mellon University Bayesian Networks (Part II) Graphical Model Readings: Murphy 10 10.2.1 Bishop 8.1,
More informationBayesian networks as causal models. Peter Antal
Bayesian networks as causal models Peter Antal antal@mit.bme.hu A.I. 3/20/2018 1 Can we represent exactly (in)dependencies by a BN? From a causal model? Suff.&nec.? Can we interpret edges as causal relations
More informationIntroduction to Artificial Intelligence. Unit # 11
Introduction to Artificial Intelligence Unit # 11 1 Course Outline Overview of Artificial Intelligence State Space Representation Search Techniques Machine Learning Logic Probabilistic Reasoning/Bayesian
More informationCS 2750: Machine Learning. Bayesian Networks. Prof. Adriana Kovashka University of Pittsburgh March 14, 2016
CS 2750: Machine Learning Bayesian Networks Prof. Adriana Kovashka University of Pittsburgh March 14, 2016 Plan for today and next week Today and next time: Bayesian networks (Bishop Sec. 8.1) Conditional
More informationMidterm, Fall 2003
5-78 Midterm, Fall 2003 YOUR ANDREW USERID IN CAPITAL LETTERS: YOUR NAME: There are 9 questions. The ninth may be more time-consuming and is worth only three points, so do not attempt 9 unless you are
More informationBayes Networks. CS540 Bryan R Gibson University of Wisconsin-Madison. Slides adapted from those used by Prof. Jerry Zhu, CS540-1
Bayes Networks CS540 Bryan R Gibson University of Wisconsin-Madison Slides adapted from those used by Prof. Jerry Zhu, CS540-1 1 / 59 Outline Joint Probability: great for inference, terrible to obtain
More informationReview of Probability Mark Craven and David Page Computer Sciences 760.
Review of Probability Mark Craven and David Page Computer Sciences 760 www.biostat.wisc.edu/~craven/cs760/ Goals for the lecture you should understand the following concepts definition of probability random
More informationSTAT 598L Probabilistic Graphical Models. Instructor: Sergey Kirshner. Probability Review
STAT 598L Probabilistic Graphical Models Instructor: Sergey Kirshner Probability Review Some slides are taken (or modified) from Carlos Guestrin s 10-708 Probabilistic Graphical Models Fall 2008 at CMU
More informationBayesian networks. Soleymani. CE417: Introduction to Artificial Intelligence Sharif University of Technology Spring 2018
Bayesian networks CE417: Introduction to Artificial Intelligence Sharif University of Technology Spring 2018 Soleymani Slides have been adopted from Klein and Abdeel, CS188, UC Berkeley. Outline Probability
More informationV7 Foundations of Probability Theory
V7 Foundations of Probability Theory Probability : degree of confidence that an event of an uncertain nature will occur. Events : we will assume that there is an agreed upon space of possible outcomes
More informationSummary of the Bayes Net Formalism. David Danks Institute for Human & Machine Cognition
Summary of the Bayes Net Formalism David Danks Institute for Human & Machine Cognition Bayesian Networks Two components: 1. Directed Acyclic Graph (DAG) G: There is a node for every variable D: Some nodes
More informationProbabilistic Graphical Models for Image Analysis - Lecture 1
Probabilistic Graphical Models for Image Analysis - Lecture 1 Alexey Gronskiy, Stefan Bauer 21 September 2018 Max Planck ETH Center for Learning Systems Overview 1. Motivation - Why Graphical Models 2.
More informationBN Semantics 3 Now it s personal! Parameter Learning 1
Readings: K&F: 3.4, 14.1, 14.2 BN Semantics 3 Now it s personal! Parameter Learning 1 Graphical Models 10708 Carlos Guestrin Carnegie Mellon University September 22 nd, 2006 1 Building BNs from independence
More informationAxioms of Probability? Notation. Bayesian Networks. Bayesian Networks. Today we ll introduce Bayesian Networks.
Bayesian Networks Today we ll introduce Bayesian Networks. This material is covered in chapters 13 and 14. Chapter 13 gives basic background on probability and Chapter 14 talks about Bayesian Networks.
More informationBayesian Networks Introduction to Machine Learning. Matt Gormley Lecture 24 April 9, 2018
10-601 Introduction to Machine Learning Machine Learning Department School of Computer Science Carnegie Mellon University Bayesian Networks Matt Gormley Lecture 24 April 9, 2018 1 Homework 7: HMMs Reminders
More informationRecall from last time. Lecture 3: Conditional independence and graph structure. Example: A Bayesian (belief) network.
ecall from last time Lecture 3: onditional independence and graph structure onditional independencies implied by a belief network Independence maps (I-maps) Factorization theorem The Bayes ball algorithm
More informationCausal Bayesian networks. Peter Antal
Causal Bayesian networks Peter Antal antal@mit.bme.hu A.I. 11/25/2015 1 Can we represent exactly (in)dependencies by a BN? From a causal model? Suff.&nec.? Can we interpret edges as causal relations with
More informationBayesian Network Representation
Bayesian Network Representation Sargur Srihari srihari@cedar.buffalo.edu 1 Topics Joint and Conditional Distributions I-Maps I-Map to Factorization Factorization to I-Map Perfect Map Knowledge Engineering
More informationBayesian networks. Instructor: Vincent Conitzer
Bayesian networks Instructor: Vincent Conitzer Rain and sprinklers example raining (X) sprinklers (Y) P(X=1) =.3 P(Y=1) =.4 grass wet (Z) Each node has a conditional probability table (CPT) P(Z=1 X=0,
More informationBayesian Networks and Decision Graphs
Bayesian Networks and Decision Graphs A 3-week course at Reykjavik University Finn V. Jensen & Uffe Kjærulff ({fvj,uk}@cs.aau.dk) Group of Machine Intelligence Department of Computer Science, Aalborg University
More informationMachine Learning CISC 5800 Dr Daniel Leeds
Machine Learning CISC 5800 Dr Daniel Leeds What is machine learning Finding patterns in data Adapting program behavior 2 Advertise a customer s favorite products This summer, I had two meetings, one in
More informationNaïve Bayes classification. p ij 11/15/16. Probability theory. Probability theory. Probability theory. X P (X = x i )=1 i. Marginal Probability
Probability theory Naïve Bayes classification Random variable: a variable whose possible values are numerical outcomes of a random phenomenon. s: A person s height, the outcome of a coin toss Distinguish
More informationProbabilistic Graphical Models (I)
Probabilistic Graphical Models (I) Hongxin Zhang zhx@cad.zju.edu.cn State Key Lab of CAD&CG, ZJU 2015-03-31 Probabilistic Graphical Models Modeling many real-world problems => a large number of random
More informationDEEP LEARNING CHAPTER 3 PROBABILITY & INFORMATION THEORY
DEEP LEARNING CHAPTER 3 PROBABILITY & INFORMATION THEORY OUTLINE 3.1 Why Probability? 3.2 Random Variables 3.3 Probability Distributions 3.4 Marginal Probability 3.5 Conditional Probability 3.6 The Chain
More informationMachine Learning, Midterm Exam: Spring 2009 SOLUTION
10-601 Machine Learning, Midterm Exam: Spring 2009 SOLUTION March 4, 2009 Please put your name at the top of the table below. If you need more room to work out your answer to a question, use the back of
More information6.047 / Computational Biology: Genomes, Networks, Evolution Fall 2008
MIT OpenCourseWare http://ocw.mit.edu 6.047 / 6.878 Computational Biology: Genomes, Networks, Evolution Fall 2008 For information about citing these materials or our Terms of Use, visit: http://ocw.mit.edu/terms.
More informationProbability and (Bayesian) Data Analysis
Department of Statistics The University of Auckland https://www.stat.auckland.ac.nz/ brewer/ Where to get everything To get all of the material (slides, code, exercises): git clone --recursive https://github.com/eggplantbren/madrid
More informationMachine Learning CISC 5800 Dr Daniel Leeds
Machine Learning CISC 5800 Dr Daniel Leeds What is machine learning Finding patterns in data Adapting program behavior 2 Dog photos and the internet Change radio channel when user says change channel Model
More informationCS 484 Data Mining. Classification 7. Some slides are from Professor Padhraic Smyth at UC Irvine
CS 484 Data Mining Classification 7 Some slides are from Professor Padhraic Smyth at UC Irvine Bayesian Belief networks Conditional independence assumption of Naïve Bayes classifier is too strong. Allows
More informationExercises, II part Exercises, II part
Inference: 12 Jul 2012 Consider the following Joint Probability Table for the three binary random variables A, B, C. Compute the following queries: 1 P(C A=T,B=T) 2 P(C A=T) P(A, B, C) A B C 0.108 T T
More informationReview: Bayesian learning and inference
Review: Bayesian learning and inference Suppose the agent has to make decisions about the value of an unobserved query variable X based on the values of an observed evidence variable E Inference problem:
More informationCausal Bayesian networks. Peter Antal
Causal Bayesian networks Peter Antal antal@mit.bme.hu A.I. 4/8/2015 1 Can we represent exactly (in)dependencies by a BN? From a causal model? Suff.&nec.? Can we interpret edges as causal relations with
More informationLecture 2: Simple Classifiers
CSC 412/2506 Winter 2018 Probabilistic Learning and Reasoning Lecture 2: Simple Classifiers Slides based on Rich Zemel s All lecture slides will be available on the course website: www.cs.toronto.edu/~jessebett/csc412
More informationProbabilistic Classification
Bayesian Networks Probabilistic Classification Goal: Gather Labeled Training Data Build/Learn a Probability Model Use the model to infer class labels for unlabeled data points Example: Spam Filtering...
More informationProbablistic Graphical Models, Spring 2007 Homework 4 Due at the beginning of class on 11/26/07
Probablistic Graphical odels, Spring 2007 Homework 4 Due at the beginning of class on 11/26/07 Instructions There are four questions in this homework. The last question involves some programming which
More informationCOMP 328: Machine Learning
COMP 328: Machine Learning Lecture 2: Naive Bayes Classifiers Nevin L. Zhang Department of Computer Science and Engineering The Hong Kong University of Science and Technology Spring 2010 Nevin L. Zhang
More informationSDS 321: Introduction to Probability and Statistics
SDS 321: Introduction to Probability and Statistics Lecture 13: Expectation and Variance and joint distributions Purnamrita Sarkar Department of Statistics and Data Science The University of Texas at Austin
More informationLecture 4 October 18th
Directed and undirected graphical models Fall 2017 Lecture 4 October 18th Lecturer: Guillaume Obozinski Scribe: In this lecture, we will assume that all random variables are discrete, to keep notations
More informationOverview of Probability. Mark Schmidt September 12, 2017
Overview of Probability Mark Schmidt September 12, 2017 Dungeons & Dragons scenario: You roll dice 1: Practical Application Roll or you sneak past monster. Otherwise, you are eaten. If you survive, you
More informationA Brief Introduction to Graphical Models. Presenter: Yijuan Lu November 12,2004
A Brief Introduction to Graphical Models Presenter: Yijuan Lu November 12,2004 References Introduction to Graphical Models, Kevin Murphy, Technical Report, May 2001 Learning in Graphical Models, Michael
More informationCS 5522: Artificial Intelligence II
CS 5522: Artificial Intelligence II Bayes Nets: Independence Instructor: Alan Ritter Ohio State University [These slides were adapted from CS188 Intro to AI at UC Berkeley. All materials available at http://ai.berkeley.edu.]
More informationProbabilistic Graphical Models
Probabilistic Graphical Models Lecture 1 Introduction CS/CNS/EE 155 Andreas Krause One of the most exciting advances in machine learning (AI, signal processing, coding, control, ) in the last decades 2
More informationUVA CS 6316/4501 Fall 2016 Machine Learning. Lecture 11: Probability Review. Dr. Yanjun Qi. University of Virginia. Department of Computer Science
UVA CS 6316/4501 Fall 2016 Machine Learning Lecture 11: Probability Review 10/17/16 Dr. Yanjun Qi University of Virginia Department of Computer Science 1 Announcements: Schedule Midterm Nov. 26 Wed / 3:30pm
More informationCOMP538: Introduction to Bayesian Networks
COMP538: Introduction to ayesian Networks Lecture 4: Inference in ayesian Networks: The VE lgorithm Nevin L. Zhang lzhang@cse.ust.hk Department of Computer Science and Engineering Hong Kong University
More informationQualifier: CS 6375 Machine Learning Spring 2015
Qualifier: CS 6375 Machine Learning Spring 2015 The exam is closed book. You are allowed to use two double-sided cheat sheets and a calculator. If you run out of room for an answer, use an additional sheet
More informationQuantifying uncertainty & Bayesian networks
Quantifying uncertainty & Bayesian networks CE417: Introduction to Artificial Intelligence Sharif University of Technology Spring 2016 Soleymani Artificial Intelligence: A Modern Approach, 3 rd Edition,
More informationBias-Variance Tradeoff
What s learning, revisited Overfitting Generative versus Discriminative Logistic Regression Machine Learning 10701/15781 Carlos Guestrin Carnegie Mellon University September 19 th, 2007 Bias-Variance Tradeoff
More informationCS 5522: Artificial Intelligence II
CS 5522: Artificial Intelligence II Bayes Nets Instructor: Alan Ritter Ohio State University [These slides were adapted from CS188 Intro to AI at UC Berkeley. All materials available at http://ai.berkeley.edu.]
More informationBN Semantics 3 Now it s personal!
Readings: K&F: 3.3, 3.4 BN Semantics 3 Now it s personal! Graphical Models 10708 Carlos Guestrin Carnegie Mellon University September 22 nd, 2008 10-708 Carlos Guestrin 2006-2008 1 Independencies encoded
More informationECE521 Tutorial 11. Topic Review. ECE521 Winter Credits to Alireza Makhzani, Alex Schwing, Rich Zemel and TAs for slides. ECE521 Tutorial 11 / 4
ECE52 Tutorial Topic Review ECE52 Winter 206 Credits to Alireza Makhzani, Alex Schwing, Rich Zemel and TAs for slides ECE52 Tutorial ECE52 Winter 206 Credits to Alireza / 4 Outline K-means, PCA 2 Bayesian
More informationReview: Directed Models (Bayes Nets)
X Review: Directed Models (Bayes Nets) Lecture 3: Undirected Graphical Models Sam Roweis January 2, 24 Semantics: x y z if z d-separates x and y d-separation: z d-separates x from y if along every undirected
More informationChapter 2: Entropy and Mutual Information. University of Illinois at Chicago ECE 534, Natasha Devroye
Chapter 2: Entropy and Mutual Information Chapter 2 outline Definitions Entropy Joint entropy, conditional entropy Relative entropy, mutual information Chain rules Jensen s inequality Log-sum inequality
More informationCS 188: Artificial Intelligence Spring Announcements
CS 188: Artificial Intelligence Spring 2011 Lecture 16: Bayes Nets IV Inference 3/28/2011 Pieter Abbeel UC Berkeley Many slides over this course adapted from Dan Klein, Stuart Russell, Andrew Moore Announcements
More information