Computer vision: models, learning and inference

Similar documents
COURSE INTRODUCTION. J. Elder CSE 6390/PSYC 6225 Computational Modeling of Visual Perception

Human-Oriented Robotics. Probability Refresher. Kai Arras Social Robotics Lab, University of Freiburg Winter term 2014/2015

BAYESIAN DECISION THEORY

A Brief Review of Probability, Bayesian Statistics, and Information Theory

Classical and Bayesian inference

Introduction to Probability

Preliminary Statistics Lecture 2: Probability Theory (Outline) prelimsoas.webs.com

Some Basic Concepts of Probability and Information Theory: Pt. 2

Notes on Probability

Terminology. Experiment = Prior = Posterior =

Communication Theory II

Some slides from Carlos Guestrin, Luke Zettlemoyer & K Gajos 2

CSC321 Lecture 18: Learning Probabilistic Models

Bayesian Inference and MCMC

Probability Based Learning

Probability Intro Part II: Bayes Rule

HST.582J / 6.555J / J Biomedical Signal and Image Processing Spring 2007

Topics. Probability Theory. Perfect Secrecy. Information Theory

Machine Learning

Machine Learning

Probability and Inference

Math 105 Course Outline

STAT 302 Introduction to Probability Learning Outcomes. Textbook: A First Course in Probability by Sheldon Ross, 8 th ed.

Introduction to Stochastic Processes

ECE521 W17 Tutorial 6. Min Bai and Yuhuai (Tony) Wu

Colorado School of Mines. Computer Vision. Professor William Hoff Dept of Electrical Engineering &Computer Science.

Data Modeling & Analysis Techniques. Probability & Statistics. Manfred Huber

Introduction to Bayesian Learning

Probabilistic Reasoning

On the errors introduced by the naive Bayes independence assumption

The devil is in the denominator

Naïve Bayes classification

Conditional probabilities and graphical models

Point Estimation. Vibhav Gogate The University of Texas at Dallas

STAT Chapter 3: Probability

Introduction to Bayesian Statistics

Bayesian RL Seminar. Chris Mansley September 9, 2008

Lecture 8. Probabilistic Reasoning CS 486/686 May 25, 2006

MACHINE LEARNING INTRODUCTION: STRING CLASSIFICATION

Language as a Stochastic Process

Conditional Probability

CS 361: Probability & Statistics

Introduction to Bayes Nets. CS 486/686: Introduction to Artificial Intelligence Fall 2013

Approximate Message Passing

PHASES OF STATISTICAL ANALYSIS 1. Initial Data Manipulation Assembling data Checks of data quality - graphical and numeric

Bayesian data analysis using JASP

Probability theory basics

Econ 325: Introduction to Empirical Economics

Basics on Probability. Jingrui He 09/11/2007

MS-A0504 First course in probability and statistics

Naïve Bayes classification. p ij 11/15/16. Probability theory. Probability theory. Probability theory. X P (X = x i )=1 i. Marginal Probability

4. Conditional Probability P( ) CSE 312 Autumn 2012 W.L. Ruzzo

Machine Learning. Bayes Basics. Marc Toussaint U Stuttgart. Bayes, probabilities, Bayes theorem & examples

Bayesian Models in Machine Learning

Bayesian Methods: Naïve Bayes

Some Probability and Statistics

Math 180B Problem Set 3

Probability and Inference. POLI 205 Doing Research in Politics. Populations and Samples. Probability. Fall 2015

Probability Rules. MATH 130, Elements of Statistics I. J. Robert Buchanan. Fall Department of Mathematics

Department of Statistical Science FIRST YEAR EXAM - SPRING 2017

9/12/17. Types of learning. Modeling data. Supervised learning: Classification. Supervised learning: Regression. Unsupervised learning: Clustering

Which coin will I use? Which coin will I use? Logistics. Statistical Learning. Topics. 573 Topics. Coin Flip. Coin Flip

Inconsistency of Bayesian inference when the model is wrong, and how to repair it

Review (probability, linear algebra) CE-717 : Machine Learning Sharif University of Technology

Probability Review Lecturer: Ji Liu Thank Jerry Zhu for sharing his slides

STA 247 Solutions to Assignment #1

Vibhav Gogate University of Texas, Dallas

LEARNING WITH BAYESIAN NETWORKS

LING 473: Day 5. START THE RECORDING Bayes Theorem. University of Washington Linguistics 473: Computational Linguistics Fundamentals

Bits. Chapter 1. Information can be learned through observation, experiment, or measurement.

CHAPTER 2 Estimating Probabilities

Probability Review. September 25, 2015

Review: Bayesian learning and inference

PERFECT SECRECY AND ADVERSARIAL INDISTINGUISHABILITY

Bayesian Analysis (Optional)

K. Nishijima. Definition and use of Bayesian probabilistic networks 1/32

(1) Introduction to Bayesian statistics

CS 188: Artificial Intelligence Spring Announcements

6.867 Machine learning, lecture 23 (Jaakkola)

Introduction to Machine Learning. Maximum Likelihood and Bayesian Inference. Lecturers: Eran Halperin, Lior Wolf

Module 2 : Conditional Probability

Machine learning: lecture 20. Tommi S. Jaakkola MIT CSAIL

Physics 403. Segev BenZvi. Classical Hypothesis Testing: The Likelihood Ratio Test. Department of Physics and Astronomy University of Rochester

Lecture 5: Bayesian Network

Hidden Markov Models. Terminology, Representation and Basic Problems

ECE521 Lecture 19 HMM cont. Inference in HMM

i=1 P ( m v) P ( v), (2) P ( m)

Quantum measurement theory

An AI-ish view of Probability, Conditional Probability & Bayes Theorem

10/18/2017. An AI-ish view of Probability, Conditional Probability & Bayes Theorem. Making decisions under uncertainty.

Markov Models and Hidden Markov Models

Machine learning: lecture 20. Tommi S. Jaakkola MIT CSAIL

Introduction to Computer Vision

Some Asymptotic Bayesian Inference (background to Chapter 2 of Tanner s book)

DEPARTMENT OF COMPUTER SCIENCE Autumn Semester MACHINE LEARNING AND ADAPTIVE INTELLIGENCE

Modeling Environment

Qualifier: CS 6375 Machine Learning Spring 2015

Random Processes. By: Nick Kingsbury

Overview of Probability. Mark Schmidt September 12, 2017

CS 188: Artificial Intelligence. Our Status in CS188

Transcription:

Computer vision: models, learning and inference Chapter 2 Introduction to probability Please send errata to s.prince@cs.ucl.ac.uk

Random variables A random variable x denotes a quantity that is uncertain May be result of experiment (flipping a coin) or a real world measurements (measuring temperature) If observe several instances of x we get different values Some values occur more than others and this information is captured by a probability distribution 2

Discrete Random Variables 3

Continuous Random Variable 4

Joint Probability Consider two random variables x and y If we observe multiple paired instances, then some combinations of outcomes are more likely than others This is captured in the joint probability distribution Written as Pr(x,y) Can read Pr(x,y) as probability of x and y 5

Joint Probability 6

Marginalization We can recover probability distribution of any variable in a joint distribution by integrating (or summing) over the other variables 7

Marginalization We can recover probability distribution of any variable in a joint distribution by integrating (or summing) over the other variables 8

Marginalization We can recover probability distribution of any variable in a joint distribution by integrating (or summing) over the other variables 9

Marginalization We can recover probability distribution of any variable in a joint distribution by integrating (or summing) over the other variables Works in higher dimensions as well leaves joint distribution between whatever variables are left 10

Conditional Probability Conditional probability of x given that y=y 1 is relative propensity of variable x to take different outcomes given that y is fixed to be equal to y 1. Written as Pr(x y=y 1 ) 11

Conditional Probability Conditional probability can be extracted from joint probability Extract appropriate slice and normalize 12

Conditional Probability More usually written in compact form Can be re-arranged to give 13

Conditional Probability This idea can be extended to more than two variables 14

Bayes Rule From before: Combining: Re-arranging: 15

Bayes Rule Terminology Likelihood propensity for observing a certain value of x given a certain value of y Prior what we know about y before seeing x Posterior what we know about y after seeing x Evidence a constant to ensure that the left hand side is a valid distribution 16

Independence If two variables x and y are independent then variable x tells us nothing about variable y (and vice-versa) 17

Independence If two variables x and y are independent then variable x tells us nothing about variable y (and vice-versa) 18

Independence When variables are independent, the joint factorizes into a product of the marginals: 19

Expectation Expectation tell us the expected or average value of some function f [x] taking into account the distribution of x Definition: 20

Expectation Expectation tell us the expected or average value of some function f [x] taking into account the distribution of x Definition in two dimensions: 21

Expectation: Common Cases 22

Expectation: Rules Rule 1: Expected value of a constant is the constant 23

Expectation: Rules Rule 2: Expected value of constant times function is constant times expected value of function 24

Expectation: Rules Rule 3: Expectation of sum of functions is sum of expectation of functions 25

Expectation: Rules Rule 4: Expectation of product of functions in variables x and y is product of expectations of functions if x and y are independent 26

Conclusions Rules of probability are compact and simple Concepts of marginalization, joint and conditional probability, Bayes rule and expectation underpin all of the models in this book One remaining concept conditional expectation discussed later 27