Probability theory: elements

Similar documents
Fusion in simple models

Artificial Intelligence

Uncertainty. Chapter 13

Artificial Intelligence Uncertainty

Uncertainty. Outline

Web-Mining Agents Data Mining

Uncertainty. Chapter 13, Sections 1 6

Probabilistic Robotics

Uncertainty. Chapter 13

Uncertainty. 22c:145 Artificial Intelligence. Problem of Logic Agents. Foundations of Probability. Axioms of Probability

Uncertainty. Outline. Probability Syntax and Semantics Inference Independence and Bayes Rule. AIMA2e Chapter 13

Uncertainty. CmpE 540 Principles of Artificial Intelligence Pınar Yolum Uncertainty. Sources of Uncertainty

Pengju XJTU 2016

COMP9414/ 9814/ 3411: Artificial Intelligence. 14. Uncertainty. Russell & Norvig, Chapter 13. UNSW c AIMA, 2004, Alan Blair, 2012

Probabilistic Reasoning

Quantifying uncertainty & Bayesian networks

Outline. Uncertainty. Methods for handling uncertainty. Uncertainty. Making decisions under uncertainty. Probability. Uncertainty

Pengju

Uncertain Knowledge and Reasoning

Uncertainty. Chapter 13. Chapter 13 1

CS 5100: Founda.ons of Ar.ficial Intelligence

CS 561: Artificial Intelligence

UNCERTAINTY. In which we see what an agent should do when not all is crystal-clear.

Chapter 13 Quantifying Uncertainty

Brief Intro. to Bayesian Networks. Extracted and modified from four lectures in Intro to AI, Spring 2008 Paul S. Rosenbloom

Cartesian-product sample spaces and independence

Basic Probability. Robert Platt Northeastern University. Some images and slides are used from: 1. AIMA 2. Chris Amato

Resolution or modus ponens are exact there is no possibility of mistake if the rules are followed exactly.

An AI-ish view of Probability, Conditional Probability & Bayes Theorem

10/18/2017. An AI-ish view of Probability, Conditional Probability & Bayes Theorem. Making decisions under uncertainty.

CS 188: Artificial Intelligence Fall 2009

Web-Mining Agents. Prof. Dr. Ralf Möller Dr. Özgür Özçep. Universität zu Lübeck Institut für Informationssysteme. Tanya Braun (Lab Class)

Probability Hal Daumé III. Computer Science University of Maryland CS 421: Introduction to Artificial Intelligence 27 Mar 2012

Basic Probability and Decisions

Reasoning with Uncertainty. Chapter 13

Uncertainty (Chapter 13, Russell & Norvig) Introduction to Artificial Intelligence CS 150 Lecture 14

Uncertainty. Russell & Norvig Chapter 13.

Our Status. We re done with Part I Search and Planning!

Probabilistic Reasoning. Kee-Eung Kim KAIST Computer Science

Probabilistic Reasoning. (Mostly using Bayesian Networks)

Artificial Intelligence CS 6364

Probabilistic Robotics. Slides from Autonomous Robots (Siegwart and Nourbaksh), Chapter 5 Probabilistic Robotics (S. Thurn et al.

Probability and Decision Theory

Probabilistic representation and reasoning

CS 188: Artificial Intelligence. Our Status in CS188

CSE 473: Artificial Intelligence

Causal Bayesian networks. Peter Antal

Probabilistic Models

Course Introduction. Probabilistic Modelling and Reasoning. Relationships between courses. Dealing with Uncertainty. Chris Williams.

PROBABILISTIC REASONING Outline

CS 687 Jana Kosecka. Uncertainty, Bayesian Networks Chapter 13, Russell and Norvig Chapter 14,

Uncertainty and Bayesian Networks

Where are we in CS 440?

Bayesian networks (1) Lirong Xia

10/15/2015 A FAST REVIEW OF DISCRETE PROBABILITY (PART 2) Probability, Conditional Probability & Bayes Rule. Discrete random variables

Reasoning under Uncertainty: Intro to Probability

13.4 INDEPENDENCE. 494 Chapter 13. Quantifying Uncertainty

Uncertainty (Chapter 13, Russell & Norvig)

Lecture Overview. Introduction to Artificial Intelligence COMP 3501 / COMP Lecture 11: Uncertainty. Uncertainty.

Probabilistic Reasoning

Causal Bayesian networks. Peter Antal

Quantifying Uncertainty & Probabilistic Reasoning. Abdulla AlKhenji Khaled AlEmadi Mohammed AlAnsari

Computer Science CPSC 322. Lecture 18 Marginalization, Conditioning

Objectives. Probabilistic Reasoning Systems. Outline. Independence. Conditional independence. Conditional independence II.

Where are we in CS 440?

Probabilistic Reasoning Systems

Bayesian networks as causal models. Peter Antal

n How to represent uncertainty in knowledge? n Which action to choose under uncertainty? q Assume the car does not have a flat tire

ARTIFICIAL INTELLIGENCE. Uncertainty: probabilistic reasoning

Random Variables. A random variable is some aspect of the world about which we (may) have uncertainty

CS 188: Artificial Intelligence Spring Announcements

Uncertainty Chapter 13. Mausam (Based on slides by UW-AI faculty)

Probability Foundations. Rudolf Kruse, Alexander Dockhorn Bayesian Networks 34

Probabilistic representation and reasoning

CS 188: Artificial Intelligence Fall 2009

CS 5522: Artificial Intelligence II

Reasoning under Uncertainty: Intro to Probability

Uncertainty. Introduction to Artificial Intelligence CS 151 Lecture 2 April 1, CS151, Spring 2004

Basic Probability and Statistics

Uncertainty. Variables. assigns to each sentence numerical degree of belief between 0 and 1. uncertainty

Why Probability? It's the right way to look at the world.

Reasoning Under Uncertainty

Bayesian networks. Soleymani. CE417: Introduction to Artificial Intelligence Sharif University of Technology Spring 2018

CS 343: Artificial Intelligence

Probabilistic Models. Models describe how (a portion of) the world works

CS 188: Artificial Intelligence Fall 2008

Vibhav Gogate University of Texas, Dallas

Uncertainty in the World. Representing Uncertainty. Uncertainty in the World and our Models. Uncertainty

Reasoning Under Uncertainty

Stochastic Methods. 5.0 Introduction 5.1 The Elements of Counting 5.2 Elements of Probability Theory

Bayesian Networks. Axioms of Probability Theory. Conditional Probability. Inference by Enumeration. Inference by Enumeration CSE 473

Consider an experiment that may have different outcomes. We are interested to know what is the probability of a particular set of outcomes.

CS188 Outline. CS 188: Artificial Intelligence. Today. Inference in Ghostbusters. Probability. We re done with Part I: Search and Planning!

Bayesian Networks. Vibhav Gogate The University of Texas at Dallas

CS 188: Artificial Intelligence Fall 2011

Probability and Uncertainty. Bayesian Networks

Bayesian classification CISC 5800 Professor Daniel Leeds

Bayesian Networks. Vibhav Gogate The University of Texas at Dallas

Graphical Models - Part I

Uncertainty. Logic and Uncertainty. Russell & Norvig. Readings: Chapter 13. One problem with logical-agent approaches: C:145 Artificial

Transcription:

Probability theory: elements Peter Antal antal@mit.bme.hu A.I. February 17, 2017 1

Joint distribution Conditional robability Indeendence, conditional indeendence Bayes rule Marginalization/Exansion Chain rule Exectation, variance

Basic element: random variable Similar to roositional logic: ossible worlds defined by assignment of values to random variables. Boolean random variables e.g., Cavity do I have a cavity? Discrete random variables e.g., Weather is one of <sunny,rainy,cloudy,snow> Domain values must be exhaustive and mutually exclusive Elementary roosition constructed by assignment of a value to a random variable: e.g., Weather = sunny, Cavity = false abbreviated as cavity Comlex roositions formed from elementary roositions and standard logical connectives e.g., Weather = sunny Cavity = false

Atomic event: A comlete secification of the state of the world about which the agent is uncertain E.g., if the world consists of only two Boolean variables Cavity and Toothache, then there are 4 distinct atomic events: Cavity = false Toothache = false Cavity = false Toothache = true Cavity = true Toothache = false Cavity = true Toothache = true Atomic events are mutually exclusive and exhaustive

For any roositions A, B 0 PA 1 Ptrue = 1 and Pfalse = 0 PA B = PA + PB - PA B

Prior or unconditional robabilities of roositions e.g., PCavity = true = 0.1 and PWeather = sunny = 0.72 corresond to belief rior to arrival of any new evidence Probability distribution gives values for all ossible assignments: PWeather = <0.72,0.1,0.08,0.1> normalized, i.e., sums to 1 Joint robability distribution for a set of random variables gives the robability of every atomic event on those random variables PWeather,Cavity = a 4 2 matrix of values: Weather = sunny rainy cloudy snow Cavity = true 0.144 0.02 0.016 0.02 Cavity = false 0.576 0.08 0.064 0.08

Conditional or osterior robabilities e.g., Pcavity toothache = 0.8 i.e., given that toothache is all I know Notation for conditional distributions: PCavity Toothache = 2-element vector of 2-element vectors If we know more, e.g., cavity is also given, then we have Pcavity toothache,cavity = 1 New evidence may be irrelevant, allowing simlification, e.g., Pcavity toothache, sunny = Pcavity toothache = 0.8 This kind of inference, sanctioned by domain knowledge, is crucial

Definition of conditional robability: Pa b = Pa b / Pb if Pb > 0 Product rule gives an alternative formulation: Pa b = Pa b Pb = Pb a Pa A general version holds for whole distributions, e.g., PWeather,Cavity = PWeather Cavity PCavity View as a set of 4 2 equations, not matrix mult. Chain rule is derived by successive alication of roduct rule: PX 1,,X n = PX 1,...,X n-1 PX n X 1,...,X n-1 = PX 1,...,X n-2 PX n-1 X 1,...,X n-2 PX n X 1,...,X n-1 = = π i= 1^n PX i X 1,,X i-1

Every question about a domain can be answered by the joint distribution. Start with the joint robability distribution: For any roosition φ, sum the atomic events where it is true: Pφ = Σ ω:ω φ Pω

Start with the joint robability distribution: For any roosition φ, sum the atomic events where it is true: Pφ = Σ ω:ω φ Pω Ptoothache = 0.108 + 0.012 + 0.016 + 0.064 = 0.2

Start with the joint robability distribution: For any roosition φ, sum the atomic events where it is true: Pφ = Σ ω:ω φ Pω Ptoothache = 0.108 + 0.012 + 0.016 + 0.064 = 0.2

Start with the joint robability distribution: Can also comute conditional robabilities: Pcavity toothache = Pcavity toothache Ptoothache = 0.016+0.064 0.108 + 0.012 + 0.016 + 0.064 = 0.4

Denominator can be viewed as a normalization constant α PCavity toothache = α, PCavity,toothache = α, [PCavity,toothache,catch + PCavity,toothache, catch] = α, [<0.108,0.016> + <0.012,0.064>] = α, <0.12,0.08> = <0.6,0.4> General idea: comute distribution on query variable by fixing evidence variables and summing over hidden variables

Tyically, we are interested in the osterior joint distribution of the query variables Y given secific values e for the evidence variables E Let the hidden variables be H = X - Y E Then the required summation of joint entries is done by summing out the hidden variables: PY E = e = αpy,e = e = ασ h PY,E= e, H = h The terms in the summation are joint entries because Y, E and H together exhaust the set of random variables Obvious roblems: 1. Worst-case time comlexity Od n where d is the largest arity 2. Sace comlexity Od n to store the joint distribution 3. How to find the numbers for Od n entries?

A and B are indeendent iff PA B = PA or PB A = PB or PA, B = PA PB PToothache, Catch, Cavity, Weather = PToothache, Catch, Cavity PWeather 32 entries reduced to 12; for n indeendent biased coins, O2 n On Absolute indeendence owerful but rare A and B are conditionally indeendent iff PA B = PA or PB A = PB or PA, B C = PA C PB C

Model Model Data Data Model X X X Y X X Y Y X X Y Y X An algebraic triviality A scientific research aradigm A ractical method for inverting causal knowledge to diagnostic tool. Cause Cause Effect Effect Cause

Probability theory=measure theory+indeendence I P X;Y Z or X Y Z P denotes that X is indeendent of Y given Z: PX;Y z=py z PX z for all z with Pz>0. Almost alternatively, I P X;Y Z iff PX Z,Y= PX Z for all z,y with Pz,y>0. Other notations: D P X;Y Z =def= I P X;Y Z Contextual indeendence: for not all z. Homeworks: Intransitivity: show that it is ossible that DX;Y, DY;Z, but IX;Z. order : show that it is ossible that IX;Z, IY;Z, but DX,Y;Z.

Joint robability distribution secifies robability of every atomic event. Queries can be answered by summing over atomic events.