Vibhav Gogate University of Texas, Dallas

Similar documents
STATISTICAL METHODS IN AI/ML Vibhav Gogate The University of Texas at Dallas. Propositional Logic and Probability Theory: Review

Bayesian Approaches Data Mining Selected Technique

Machine Learning

Consider an experiment that may have different outcomes. We are interested to know what is the probability of a particular set of outcomes.

CS37300 Class Notes. Jennifer Neville, Sebastian Moreno, Bruno Ribeiro

Econ 325: Introduction to Empirical Economics

Machine Learning

Uncertainty. Chapter 13

Basic Probability. Robert Platt Northeastern University. Some images and slides are used from: 1. AIMA 2. Chris Amato

{ p if x = 1 1 p if x = 0

STAT:5100 (22S:193) Statistical Inference I

Resolution or modus ponens are exact there is no possibility of mistake if the rules are followed exactly.

CS626 Data Analysis and Simulation

Lecture 2: Probability, conditional probability, and independence

Uncertainty and Bayesian Networks

Why Probability? It's the right way to look at the world.

Data Modeling & Analysis Techniques. Probability & Statistics. Manfred Huber

Computer Science CPSC 322. Lecture 18 Marginalization, Conditioning

Basic Probability and Statistics

Probability Theory for Machine Learning. Chris Cremer September 2015

An AI-ish view of Probability, Conditional Probability & Bayes Theorem

10/18/2017. An AI-ish view of Probability, Conditional Probability & Bayes Theorem. Making decisions under uncertainty.

Uncertainty. Logic and Uncertainty. Russell & Norvig. Readings: Chapter 13. One problem with logical-agent approaches: C:145 Artificial

Uncertain Knowledge and Bayes Rule. George Konidaris

Unit 7 Probability M2 13.1,2,4, 5,6

Lecture 3 - Axioms of Probability

Cartesian-product sample spaces and independence

What is Probability? Probability. Sample Spaces and Events. Simple Event

COMP 328: Machine Learning

Why study probability? Set theory. ECE 6010 Lecture 1 Introduction; Review of Random Variables

Single Maths B: Introduction to Probability

Rapid Introduction to Machine Learning/ Deep Learning

Probability theory: elements

Lecture 11. Probability Theory: an Overveiw

Probability theory. References:

Machine Learning

Probability. Paul Schrimpf. January 23, UBC Economics 326. Probability. Paul Schrimpf. Definitions. Properties. Random variables.

Review (Probability & Linear Algebra)

Lecture 10: Introduction to reasoning under uncertainty. Uncertainty

Naïve Bayes. Jia-Bin Huang. Virginia Tech Spring 2019 ECE-5424G / CS-5824

Uncertainty (Chapter 13, Russell & Norvig) Introduction to Artificial Intelligence CS 150 Lecture 14

Artificial Intelligence

Probability Review. Yutian Li. January 18, Stanford University. Yutian Li (Stanford University) Probability Review January 18, / 27

Generative Techniques: Bayes Rule and the Axioms of Probability

Some Probability and Statistics

Uncertainty. Outline

Priors, Total Probability, Expectation, Multiple Trials

Probability Theory and Applications

Lecture 2. Conditional Probability

Probability Review Lecturer: Ji Liu Thank Jerry Zhu for sharing his slides

Probability calculus and statistics

Stochastic Methods. 5.0 Introduction 5.1 The Elements of Counting 5.2 Elements of Probability Theory

Introduction to Bayes Nets. CS 486/686: Introduction to Artificial Intelligence Fall 2013

Brief Review of Probability

Topic 2 Probability. Basic probability Conditional probability and independence Bayes rule Basic reliability

Review (probability, linear algebra) CE-717 : Machine Learning Sharif University of Technology

Probability. Paul Schrimpf. January 23, Definitions 2. 2 Properties 3

COMP9414/ 9814/ 3411: Artificial Intelligence. 14. Uncertainty. Russell & Norvig, Chapter 13. UNSW c AIMA, 2004, Alan Blair, 2012

Dept. of Linguistics, Indiana University Fall 2015

Mean, Median and Mode. Lecture 3 - Axioms of Probability. Where do they come from? Graphically. We start with a set of 21 numbers, Sta102 / BME102

CS 5100: Founda.ons of Ar.ficial Intelligence

Classification and Regression Trees

Human-Oriented Robotics. Probability Refresher. Kai Arras Social Robotics Lab, University of Freiburg Winter term 2014/2015

Basic Probabilistic Reasoning SEG

Probabilistic Reasoning

Determining Probabilities. Product Rule for Ordered Pairs/k-Tuples:

Basic notions of probability theory

Where are we in CS 440?

Uncertainty. Chapter 13, Sections 1 6

Introduction to Machine Learning

Introduction to Probability and Stocastic Processes - Part I

Artificial Intelligence Uncertainty

Uncertainty. Chapter 13

Introduction to Stochastic Processes

CS 331: Artificial Intelligence Probability I. Dealing with Uncertainty

Review of probability

Basic notions of probability theory

The Naïve Bayes Classifier. Machine Learning Fall 2017

7.1 What is it and why should we care?

Mobile Robotics II: Simultaneous localization and mapping

Lecture 9: Naive Bayes, SVM, Kernels. Saravanan Thirumuruganathan

Uncertainty. 22c:145 Artificial Intelligence. Problem of Logic Agents. Foundations of Probability. Axioms of Probability

Information Science 2

Lecture 04: Conditional Probability. Lisa Yan July 2, 2018

Overview of Probability. Mark Schmidt September 12, 2017

Lecture 4 : Conditional Probability and Bayes Theorem 0/ 26

Module 2 : Conditional Probability

COMP61011 : Machine Learning. Probabilis*c Models + Bayes Theorem

Machine Learning. Yuh-Jye Lee. March 1, Lab of Data Science and Machine Intelligence Dept. of Applied Math. at NCTU

Recitation 2: Probability

Notes on Probability

COMP5211 Lecture Note on Reasoning under Uncertainty

Test One Mathematics Fall 2009

Decision Trees. Nicholas Ruozzi University of Texas at Dallas. Based on the slides of Vibhav Gogate and David Sontag

Probabilistic Reasoning. (Mostly using Bayesian Networks)

Part (A): Review of Probability [Statistics I revision]

3.1 Events, Sample Spaces, and Probability

Basic Probability and Decisions

Exam 1. Problem 1: True or false

STAT Chapter 3: Probability

Transcription:

Review of Probability and Statistics 101

Elements of Probability Theory Events, Sample Space and Random Variables Axioms of Probability Independent Events Conditional Probability Bayes Theorem Joint Probability Distribution Expectations and Variance Independence and Conditional Independence Continuous versus Discrete Distributions Common Continuous and Discrete Distributions

Events, Sample Space and Random Variables A sample space is a set of possible outcomes in your domain. All possible entries in a truth table. Can be Infinite. Example: Set of Real numbers Random Variable is a function defined over the sample space S A Boolean random variable X: S {True, False} Stock price of Google G: S Set of Reals An Event is a subset of S A subset of S for which X = True. Stock price of Google is between 575 and 580.

Events, Sample Space and Random Variables: Picture Sample Space: The Rectangle. Random variable: A. Event: A is True Probability: A real function defined over the events in the sample space.

Axioms of Probability Four Axioms of Probability: 0 P(A) 1 P(True) = 1 (i.e., an event in which all outcomes occur) P(False) = 0 (i.e., an event in no outcomes occur) P(A B) = P(A) + P(B) P(A B)

Probability Densities Probability Density: p(x (a, b)) = b a p(x)dx Cumulative Distribution Function: P(z) = z p(x)dx Such that: p(x) 0 p(x)dx = 1

Probability Mass Functions A 1,..., A n is a set of mutually exclusive events such that n P(A i ) = 1 i=1 P is called a probability mass function or a probability distribution. Each A i can be regarded as specific value in the discretization of a continuous quantity.

Sum Rule 0 P(A) 1 P(True) = 1 (i.e., an event in which all outcomes occur) P(False) = 0 (i.e., an event in no outcomes occur) P(A B) = P(A) + P(B) P(A B) To prove that: 1 P(A) = 1 P( A) 2 P(A) = P(A B) + P(A B) SUM RULE: P(A) = n P(A B i ) i=1 where {B 1,..., B n } is a set of of mutually exclusive and exhaustive events.

Conditional Probability P(A B) = P(A B) P(B)

Chain Rule P(A B) = P(A B) P(B) P(A B) = P(A B)P(B) P(A B C) = P(A B C)P(B C)P(C) n P(A 1 A 2... A n ) = P(A i A 1... A i 1 ) i=1

Independence and Conditional Independence Independence: Two events are independent if P(A B) = P(A)P(B) Implies that: P(A B) = P(A) and P(B A) = P(B) Knowing A tells me nothing about B and vice versa. A: Getting a 3 on the face of a die. B: New England Patriots win the Superbowl. Conditional Independence: A and C are conditionally independent given B iff P(A B C) = P(A B) Knowing C tells us nothing about A given B.

Bayes Rule P(A B) = P(B A)P(A) P(B) Proof. P(A B) = P(A B) P(B) (1) P(B A) = P(A B) P(A) (2) Therefore, P(A B) = P(B A)P(A) (3) Substituting P(A B) in Equation (1), we get Bayes Rule.

Other Forms of Bayes Rule Form 1: P(A B) = = P(B A)P(A) P(A B) + P( A B) P(B A)P(A) P(B A)P(A) + P(B A)P( A) (1) (2) Form 2: P(A B C) = P(B A C)P(A C) P(B C)

Applying Bayes Rule: Example The probability that a person fails a lie detector test given that he/she is cheating on his/her partner is 0.98. The probability that a person fails the test given that he/she is not cheating on his/her partner is 0.05. You are a CS graduate student and the probability that a CS graduate student will cheat on his/her partner is 1 in 10000 (CS grads are boring!). A person will break up with his/her partner if the probability that the partner is cheating is greater than 0.005 (i.e., > 0.5%). Today, you come home and you find out that you have failed the lie detector test. Convince him/her that he/she should not break up with you.

Another Interpretation of the Bayes Rule posterior = likelihood prior Probability of evidence P(Test = Fail Cheating = yes) P(Cheating = yes) P(Cheating = yes Test = Fail) = P(Test = Fail) Prior probability of cheating Likelihood of failing the test given that a person is cheating Test=Fail is the evidence

Expectation and Variance Expectation: E[f ] = p(x)f (x) x E[f ] = p(x)f (x)dx Conditional Expectation: E[f y] = x p(x y)f (x) Variance: var[f ] = E[f (x) 2 ] E[f (x)] 2

Joint Distribution Assign a probability value to joint assignments to random variables. If all variables are discrete, we consider Cartesian product of their sets of values For Boolean variables, we attach a value to each row of a truth table The sum of probabilities should sum to 1. Outlook Humidity Tennis? Value Sunny High Yes 0.05 Sunny High No 0.2 Sunny Normal Yes 0.2 Sunny Normal No 0.1 Windy High Yes 0.2 Windy High No 0.05 Windy Normal Yes 0.05 Windy Normal No 0.15

The Joint Distribution Represents complete knowledge about the domain Can be used to answer any question that you might have about the domain P(Event) = Sum of Probabilities where the Event is True P(Outlook = Sunny) = P(Humidity = High Tennis? = No) = P(Humidity = High Tennis? = No) = Outlook Humidity Tennis? Value Sunny High Yes 0.05 Sunny High No 0.2 Sunny Normal Yes 0.2 Sunny Normal No 0.1 Windy High Yes 0.2 Windy High No 0.05 Windy Normal Yes 0.05 Windy Normal No 0.15