Colorado School of Mines. Computer Vision. Professor William Hoff Dept of Electrical Engineering &Computer Science.
|
|
- Elwin Eric Long
- 5 years ago
- Views:
Transcription
1 rofessor William Hoff Dept of Electrical Engineering &Computer Science 1
2 Review of robability For additional review material, see 02 Reviews/Reviewrobability.pdf Note some of the figures in this lecture are from Computer vision: models, learning and inference, by Simon J.D. rince, Cambridge University ress 2012
3 robability robability Marginalization Bayes rule Expectation Fitting probability models Robust statistics 3
4 robability The probability of an event is denoted as is the outcome of a random experiment we call this a random variable If we do a large number of experiments, the relative frequency of occurring approaches is between 0 and 1 inclusive is the joint probability of both events and occurring We can also write this as, The conditional probability of event that event has occurred, is occurring, given 4
5 Example ull a card at random from a deck of cards What is the conditional probability that the card is the ace of clubs, given that it is a black card? Solution: A is the event that it is the ace of clubs B is the event that it is a black card 1/52 26/ we know that AB=A because if the card is the ace of clubs, it must be black 5
6 Marginalization Let,,,, be an exhaustive, mutually exclusive set of events We can find the probability of another event by summing over conditional probabilities this is called the marginal probability : 6
7 Example A company makes widgets from three machines Machine M1 makes 3000/hour, and 80% are good Machine M2 makes 4000/hour, and 90% are good Machine M3 makes 3000/hour, and 60% are good All widgets are mixed together. What is the probability that a widget drawn at random is good? 7
8 Solution Let A be the event that the widget is good We know 0.3, 0.4, , 0.9, 0.6 By marginalization, = =
9 Bayes Rule Conditional probability relationships: The second expression may be written as which is known as Bayes' theorem, so named after the 18th century mathematician Thomas Bayes. 9 B AB B A A A B B B A AB A B B A A B
10 Example A blood test can test for the presence of a disease. Joe's doctor draws some of Joe's blood, and performs the test on his blood. The results indicate that the disease is present in Joe. Here's the information that Joe's doctor knows about the disease and the diagnostic blood test: 1 in 100 people have the disease. That is, if D is the event that a randomly selected individual has the disease, then pd = If H is the event that a randomly selected individual is healthy, then ph = If a person has the disease, then the probability that the blood test comes back positive is That is, pt+ D = If a person is healthy, then the probability that the diagnostic test comes back negative is That is, pt H = What is the probability that Joe has the disease? 10
11 Solution 11
12 Continuous variables A random variable can be continuous, taking on any real number The probability density function pdf describes the probability of a continuous variable Instead of sums, we use integrals From Computer vision: models, learning and inference, by Simon J.D. rince, Cambridge University ress
13 Expectation The expected value, or expectation, of a random variable is just the average We can compute it by summing or integrating over probabilities We can also compute the expectation of a function of a random variable 13
14 Fitting robability Models Maximum Likelihood Maximum likelihood ML method: find the parameters under which the data are most likely for that model: Example: the normal Gaussian probability model 14
15 Least Squares Fitting Assumes that measurements follow a normal distribution To find the parameters with the maximum likelihood, we use the log of the probability called the log likelihood These are constants So we are just minimizing the squared error 15
16 Least Squares Fitting continued To maximize, differentiate the log likelihood with respect to the parameters and set to zero: We get And similarly for 16
17 Fitting robability Models Maximum a osteriori Maximum a posteriori MA method: find the parameters to maximize the probability of the parameters given the data: This method is good to use when we have some prior information about the parameters We can discard the denominator as it is constant with respect to the parameters and so does not affect the position of the maximum 17
18 Robust Statistics What if the noise is not Gaussian? Measurements can be contaminated with large errors; e.g., due to gross failures in the measurement process Since residual errors are squared, a single large error can have a large influence on the fit In such cases, you can use a penalty function for the residuals that has a slower growth instead of squaring them This formulation of the inference problem is called an Mestimator in the robust statistics literature See appendix B.3 in the Szeliski book 18
19 Iteratively Re weighted Least Squares We can take the derivative of with respect to p and set it to 0: where is the derivative of ρ and is called the influence function. If we introduce a weight function, equivalent to minimizing, solving the above is The Iteratively re weighted least squares IRLS algorithm alternates between computing the weight functions and solving the resulting weighted least squares problem with fixed w values You should start IRLS in the vicinity of a good solution using for example, RANSAC 19
20 From us/um/people/zhang/inria/ublis/tutorial Estim/node24.html 20
21 From us/um/people/zhang/inria/ublis/tutorial Estim/node24.html 21
Slide 1 Math 1520, Lecture 21
Slide 1 Math 1520, Lecture 21 This lecture is concerned with a posteriori probability, which is the probability that a previous event had occurred given the outcome of a later event. Slide 2 Conditional
More informationStatistics for Business and Economics
Statistics for Business and Economics Basic Probability Learning Objectives In this lecture(s), you learn: Basic probability concepts Conditional probability To use Bayes Theorem to revise probabilities
More informationENGR 200 ENGR 200. What did we do last week?
ENGR 200 What did we do last week? Definition of probability xioms of probability Sample space robability laws Conditional probability ENGR 200 Lecture 3: genda. Conditional probability 2. Multiplication
More informationReview of probabilities
CS 1675 Introduction to Machine Learning Lecture 5 Density estimation Milos Hauskrecht milos@pitt.edu 5329 Sennott Square Review of probabilities 1 robability theory Studies and describes random processes
More informationProbability and Probability Distributions. Dr. Mohammed Alahmed
Probability and Probability Distributions 1 Probability and Probability Distributions Usually we want to do more with data than just describing them! We might want to test certain specific inferences about
More informationProbability. Introduction to Biostatistics
Introduction to Biostatistics Probability Second Semester 2014/2015 Text Book: Basic Concepts and Methodology for the Health Sciences By Wayne W. Daniel, 10 th edition Dr. Sireen Alkhaldi, BDS, MPH, DrPH
More informationConditional Probability. CS231 Dianna Xu
Conditional Probability CS231 Dianna Xu 1 Boy or Girl? A couple has two children, one of them is a girl. What is the probability that the other one is also a girl? Assuming 50/50 chances of conceiving
More information4. Probability of an event A for equally likely outcomes:
University of California, Los Angeles Department of Statistics Statistics 110A Instructor: Nicolas Christou Probability Probability: A measure of the chance that something will occur. 1. Random experiment:
More informationSTAT 516: Basic Probability and its Applications
Lecture 3: Conditional Probability and Independence Prof. Michael September 29, 2015 Motivating Example Experiment ξ consists of rolling a fair die twice; A = { the first roll is 6 } amd B = { the sum
More informationSTAT Chapter 3: Probability
Basic Definitions STAT 515 --- Chapter 3: Probability Experiment: A process which leads to a single outcome (called a sample point) that cannot be predicted with certainty. Sample Space (of an experiment):
More information2) There should be uncertainty as to which outcome will occur before the procedure takes place.
robability Numbers For many statisticians the concept of the probability that an event occurs is ultimately rooted in the interpretation of an event as an outcome of an experiment, others would interpret
More informationEcon 325: Introduction to Empirical Economics
Econ 325: Introduction to Empirical Economics Lecture 2 Probability Copyright 2010 Pearson Education, Inc. Publishing as Prentice Hall Ch. 3-1 3.1 Definition Random Experiment a process leading to an uncertain
More information2011 Pearson Education, Inc
Statistics for Business and Economics Chapter 3 Probability Contents 1. Events, Sample Spaces, and Probability 2. Unions and Intersections 3. Complementary Events 4. The Additive Rule and Mutually Exclusive
More informationELEG 3143 Probability & Stochastic Process Ch. 1 Probability
Department of Electrical Engineering University of Arkansas ELEG 3143 Probability & Stochastic Process Ch. 1 Probability Dr. Jingxian Wu wuj@uark.edu OUTLINE 2 Applications Elementary Set Theory Random
More informationIntroduction to Bayesian Statistics and Markov Chain Monte Carlo Estimation. EPSY 905: Multivariate Analysis Spring 2016 Lecture #10: April 6, 2016
Introduction to Bayesian Statistics and Markov Chain Monte Carlo Estimation EPSY 905: Multivariate Analysis Spring 2016 Lecture #10: April 6, 2016 EPSY 905: Intro to Bayesian and MCMC Today s Class An
More informationProbability and statistics; Rehearsal for pattern recognition
Probability and statistics; Rehearsal for pattern recognition Václav Hlaváč Czech Technical University in Prague Czech Institute of Informatics, Robotics and Cybernetics 166 36 Prague 6, Jugoslávských
More informationOrigins of Probability Theory
1 16.584: INTRODUCTION Theory and Tools of Probability required to analyze and design systems subject to uncertain outcomes/unpredictability/randomness. Such systems more generally referred to as Experiments.
More informationIntroduction to Bayesian Learning. Machine Learning Fall 2018
Introduction to Bayesian Learning Machine Learning Fall 2018 1 What we have seen so far What does it mean to learn? Mistake-driven learning Learning by counting (and bounding) number of mistakes PAC learnability
More informationMachine Learning. Lecture 4: Regularization and Bayesian Statistics. Feng Li. https://funglee.github.io
Machine Learning Lecture 4: Regularization and Bayesian Statistics Feng Li fli@sdu.edu.cn https://funglee.github.io School of Computer Science and Technology Shandong University Fall 207 Overfitting Problem
More informationIntro to Probability Day 3 (Compound events & their probabilities)
Intro to Probability Day 3 (Compound events & their probabilities) Compound Events Let A, and B be two event. Then we can define 3 new events as follows: 1) A or B (also A B ) is the list of all outcomes
More informationProperties of Probability
Econ 325 Notes on Probability 1 By Hiro Kasahara Properties of Probability In statistics, we consider random experiments, experiments for which the outcome is random, i.e., cannot be predicted with certainty.
More informationLecture : Probabilistic Machine Learning
Lecture : Probabilistic Machine Learning Riashat Islam Reasoning and Learning Lab McGill University September 11, 2018 ML : Many Methods with Many Links Modelling Views of Machine Learning Machine Learning
More informationCSC321 Lecture 18: Learning Probabilistic Models
CSC321 Lecture 18: Learning Probabilistic Models Roger Grosse Roger Grosse CSC321 Lecture 18: Learning Probabilistic Models 1 / 25 Overview So far in this course: mainly supervised learning Language modeling
More informationSparse Linear Models (10/7/13)
STA56: Probabilistic machine learning Sparse Linear Models (0/7/) Lecturer: Barbara Engelhardt Scribes: Jiaji Huang, Xin Jiang, Albert Oh Sparsity Sparsity has been a hot topic in statistics and machine
More informationProbability Theory for Machine Learning. Chris Cremer September 2015
Probability Theory for Machine Learning Chris Cremer September 2015 Outline Motivation Probability Definitions and Rules Probability Distributions MLE for Gaussian Parameter Estimation MLE and Least Squares
More informationIntro. ANN & Fuzzy Systems. Lecture 15. Pattern Classification (I): Statistical Formulation
Lecture 15. Pattern Classification (I): Statistical Formulation Outline Statistical Pattern Recognition Maximum Posterior Probability (MAP) Classifier Maximum Likelihood (ML) Classifier K-Nearest Neighbor
More informationComputer vision: models, learning and inference
Computer vision: models, learning and inference Chapter 2 Introduction to probability Please send errata to s.prince@cs.ucl.ac.uk Random variables A random variable x denotes a quantity that is uncertain
More informationProblem # Number of points 1 /20 2 /20 3 /20 4 /20 5 /20 6 /20 7 /20 8 /20 Total /150
Name Student ID # Instructor: SOLUTION Sergey Kirshner STAT 516 Fall 09 Practice Midterm #1 January 31, 2010 You are not allowed to use books or notes. Non-programmable non-graphic calculators are permitted.
More informationLecture 4: Probabilistic Learning. Estimation Theory. Classification with Probability Distributions
DD2431 Autumn, 2014 1 2 3 Classification with Probability Distributions Estimation Theory Classification in the last lecture we assumed we new: P(y) Prior P(x y) Lielihood x2 x features y {ω 1,..., ω K
More information2 Chapter 2: Conditional Probability
STAT 421 Lecture Notes 18 2 Chapter 2: Conditional Probability Consider a sample space S and two events A and B. For example, suppose that the equally likely sample space is S = {0, 1, 2,..., 99} and A
More informationBayesian Deep Learning
Bayesian Deep Learning Mohammad Emtiyaz Khan AIP (RIKEN), Tokyo http://emtiyaz.github.io emtiyaz.khan@riken.jp June 06, 2018 Mohammad Emtiyaz Khan 2018 1 What will you learn? Why is Bayesian inference
More informationPLEASE MARK YOUR ANSWERS WITH AN X, not a circle! 1. (a) (b) (c) (d) (e) 2. (a) (b) (c) (d) (e) (a) (b) (c) (d) (e) 4. (a) (b) (c) (d) (e)...
Math 020, Exam II October, 206 The Honor Code is in effect for this examination. All work is to be your own. You may use a calculator. The exam lasts for hour 5 minutes. Be sure that your name is on every
More information2.4. Conditional Probability
2.4. Conditional Probability Objectives. Definition of conditional probability and multiplication rule Total probability Bayes Theorem Example 2.4.1. (#46 p.80 textbook) Suppose an individual is randomly
More informationCHAPTER 4. Probability is used in inference statistics as a tool to make statement for population from sample information.
CHAPTER 4 PROBABILITY Probability is used in inference statistics as a tool to make statement for population from sample information. Experiment is a process for generating observations Sample space is
More informationLecture 4: Probabilistic Learning
DD2431 Autumn, 2015 1 Maximum Likelihood Methods Maximum A Posteriori Methods Bayesian methods 2 Classification vs Clustering Heuristic Example: K-means Expectation Maximization 3 Maximum Likelihood Methods
More informationParameter Estimation. Industrial AI Lab.
Parameter Estimation Industrial AI Lab. Generative Model X Y w y = ω T x + ε ε~n(0, σ 2 ) σ 2 2 Maximum Likelihood Estimation (MLE) Estimate parameters θ ω, σ 2 given a generative model Given observed
More informationMath Exam 1 Review. NOTE: For reviews of the other sections on Exam 1, refer to the first page of WIR #1 and #2.
Math 166 Fall 2008 c Heather Ramsey Page 1 Math 166 - Exam 1 Review NOTE: For reviews of the other sections on Exam 1, refer to the first page of WIR #1 and #2. Section 1.5 - Rules for Probability Elementary
More informationMachine Learning. Topic 12: Probability and Bayesian Networks
Machine Learning Topic 12: robability and ayesian Networks ryan ardo, Machine Learning: EECS 349 Fall 2009 xioms of robability Let there be a space S composed of a countable number of events The probability
More informationLecture 4. Selected material from: Ch. 6 Probability
Lecture 4 Selected material from: Ch. 6 Probability Example: Music preferences F M Suppose you want to know what types of CD s males and females are more likely to buy. The CD s are classified as Classical,
More informationMathematics 102 Solutions for HWK 22 SECTION 7.6 P (R 1 )=0.05
Mathematics 102 Solutions for HWK 22 SECTION 7.6 p 368 Problem 3. Assume R 1, R 2, and R 3 are mutually exclusive events and we have P (R 1 )0.05 P (R 2 )0.6 P (R 3 )0.35 P (Q R 1 )0.4 P (Q R 2 )0.3 P
More informationMixture Models and EM
Mixture Models and EM Goal: Introduction to probabilistic mixture models and the expectationmaximization (EM) algorithm. Motivation: simultaneous fitting of multiple model instances unsupervised clustering
More informationChapter 7: Section 7-1 Probability Theory and Counting Principles
Chapter 7: Section 7-1 Probability Theory and Counting Principles D. S. Malik Creighton University, Omaha, NE D. S. Malik Creighton University, Omaha, NE Chapter () 7: Section 7-1 Probability Theory and
More informationStatistics for Managers Using Microsoft Excel (3 rd Edition)
Statistics for Managers Using Microsoft Excel (3 rd Edition) Chapter 4 Basic Probability and Discrete Probability Distributions 2002 Prentice-Hall, Inc. Chap 4-1 Chapter Topics Basic probability concepts
More informationSome Concepts of Probability (Review) Volker Tresp Summer 2018
Some Concepts of Probability (Review) Volker Tresp Summer 2018 1 Definition There are different way to define what a probability stands for Mathematically, the most rigorous definition is based on Kolmogorov
More informationNaïve Bayes classification
Naïve Bayes classification 1 Probability theory Random variable: a variable whose possible values are numerical outcomes of a random phenomenon. Examples: A person s height, the outcome of a coin toss
More informationExample. What is the sample space for flipping a fair coin? Rolling a 6-sided die? Find the event E where E = {x x has exactly one head}
Chapter 7 Notes 1 (c) Epstein, 2013 CHAPTER 7: PROBABILITY 7.1: Experiments, Sample Spaces and Events Chapter 7 Notes 2 (c) Epstein, 2013 What is the sample space for flipping a fair coin three times?
More informationBayesian Learning Features of Bayesian learning methods:
Bayesian Learning Features of Bayesian learning methods: Each observed training example can incrementally decrease or increase the estimated probability that a hypothesis is correct. This provides a more
More informationINTRODUCTION TO PATTERN RECOGNITION
INTRODUCTION TO PATTERN RECOGNITION INSTRUCTOR: WEI DING 1 Pattern Recognition Automatic discovery of regularities in data through the use of computer algorithms With the use of these regularities to take
More informationBayesian Linear Regression [DRAFT - In Progress]
Bayesian Linear Regression [DRAFT - In Progress] David S. Rosenberg Abstract Here we develop some basics of Bayesian linear regression. Most of the calculations for this document come from the basic theory
More informationHuman-Oriented Robotics. Probability Refresher. Kai Arras Social Robotics Lab, University of Freiburg Winter term 2014/2015
Probability Refresher Kai Arras, University of Freiburg Winter term 2014/2015 Probability Refresher Introduction to Probability Random variables Joint distribution Marginalization Conditional probability
More informationProbability 5-4 The Multiplication Rules and Conditional Probability
Outline Lecture 8 5-1 Introduction 5-2 Sample Spaces and 5-3 The Addition Rules for 5-4 The Multiplication Rules and Conditional 5-11 Introduction 5-11 Introduction as a general concept can be defined
More informationProbability and Information Theory. Sargur N. Srihari
Probability and Information Theory Sargur N. srihari@cedar.buffalo.edu 1 Topics in Probability and Information Theory Overview 1. Why Probability? 2. Random Variables 3. Probability Distributions 4. Marginal
More informationConditional Probability
Example 2.24 Complex components are assembled in a plant that uses two different assembly lines, A and B. Line A uses older equipment than B, so it is somewhat slower and less reliable. Suppose on a given
More information7.1 What is it and why should we care?
Chapter 7 Probability In this section, we go over some simple concepts from probability theory. We integrate these with ideas from formal language theory in the next chapter. 7.1 What is it and why should
More informationMODULE -4 BAYEIAN LEARNING
MODULE -4 BAYEIAN LEARNING CONTENT Introduction Bayes theorem Bayes theorem and concept learning Maximum likelihood and Least Squared Error Hypothesis Maximum likelihood Hypotheses for predicting probabilities
More informationReview of Statistics
Review of Statistics Topics Descriptive Statistics Mean, Variance Probability Union event, joint event Random Variables Discrete and Continuous Distributions, Moments Two Random Variables Covariance and
More informationRelationship between Least Squares Approximation and Maximum Likelihood Hypotheses
Relationship between Least Squares Approximation and Maximum Likelihood Hypotheses Steven Bergner, Chris Demwell Lecture notes for Cmpt 882 Machine Learning February 19, 2004 Abstract In these notes, a
More informationFrom Bayes Theorem to Pattern Recognition via Bayes Rule
From Bayes Theorem to Pattern Recognition via Bayes Rule Slecture by Varun Vasudevan (partially based on Prof. Mireille Boutin s ECE 662 lecture) February 12, 2014 What will you learn from this slecture?
More informationtossing a coin selecting a card from a deck measuring the commuting time on a particular morning
2 Probability Experiment An experiment or random variable is any activity whose outcome is unknown or random upfront: tossing a coin selecting a card from a deck measuring the commuting time on a particular
More informationClustering and Gaussian Mixture Models
Clustering and Gaussian Mixture Models Piyush Rai IIT Kanpur Probabilistic Machine Learning (CS772A) Jan 25, 2016 Probabilistic Machine Learning (CS772A) Clustering and Gaussian Mixture Models 1 Recap
More informationElementary Discrete Probability
Elementary Discrete Probability MATH 472 Financial Mathematics J Robert Buchanan 2018 Objectives In this lesson we will learn: the terminology of elementary probability, elementary rules of probability,
More informationSTAT:5100 (22S:193) Statistical Inference I
STAT:5100 (22S:193) Statistical Inference I Week 3 Luke Tierney University of Iowa Fall 2015 Luke Tierney (U Iowa) STAT:5100 (22S:193) Statistical Inference I Fall 2015 1 Recap Matching problem Generalized
More informationSTAT 414: Introduction to Probability Theory
STAT 414: Introduction to Probability Theory Spring 2016; Homework Assignments Latest updated on April 29, 2016 HW1 (Due on Jan. 21) Chapter 1 Problems 1, 8, 9, 10, 11, 18, 19, 26, 28, 30 Theoretical Exercises
More informationA survey of Probability concepts. Chapter 5
A survey of Probability concepts Chapter 5 Learning Objectives Define probability. Explain the terms experiment, event, outcome. Define the terms conditional probability and joint probability. Calculate
More informationDiscrete Mathematics and Probability Theory Spring 2016 Rao and Walrand Note 14
CS 70 Discrete Mathematics and Probability Theory Spring 2016 Rao and Walrand Note 14 Introduction One of the key properties of coin flips is independence: if you flip a fair coin ten times and get ten
More informationStatistical and Learning Techniques in Computer Vision Lecture 2: Maximum Likelihood and Bayesian Estimation Jens Rittscher and Chuck Stewart
Statistical and Learning Techniques in Computer Vision Lecture 2: Maximum Likelihood and Bayesian Estimation Jens Rittscher and Chuck Stewart 1 Motivation and Problem In Lecture 1 we briefly saw how histograms
More informationRelative Risks (RR) and Odds Ratios (OR) 20
BSTT523: Pagano & Gavreau, Chapter 6 1 Chapter 6: Probability slide: Definitions (6.1 in P&G) 2 Experiments; trials; probabilities Event operations 4 Intersection; Union; Complement Venn diagrams Conditional
More information3 PROBABILITY TOPICS
Chapter 3 Probability Topics 135 3 PROBABILITY TOPICS Figure 3.1 Meteor showers are rare, but the probability of them occurring can be calculated. (credit: Navicore/flickr) Introduction It is often necessary
More informationPreliminary Statistics Lecture 2: Probability Theory (Outline) prelimsoas.webs.com
1 School of Oriental and African Studies September 2015 Department of Economics Preliminary Statistics Lecture 2: Probability Theory (Outline) prelimsoas.webs.com Gujarati D. Basic Econometrics, Appendix
More informationBasics of Bayesian analysis. Jorge Lillo-Box MCMC Coffee Season 1, Episode 5
Basics of Bayesian analysis Jorge Lillo-Box MCMC Coffee Season 1, Episode 5 Disclaimer This presentation is based heavily on Statistics, Data Mining, and Machine Learning in Astronomy: A Practical Python
More informationConditional Probability and Independence
Conditional Probability and Independence September 3, 2009 1 Restricting the Sample Space - Conditional Probability How do we modify the probability of an event in light of the fact that something is known?
More informationMachine Learning 4. week
Machine Learning 4. week robability and Conditional robability ayes Theorem Naïve ayes Classifier Umut ORHN, hd. robability The term shows the occurring likelihood of each situation in a random process
More informationNaïve Bayes classification. p ij 11/15/16. Probability theory. Probability theory. Probability theory. X P (X = x i )=1 i. Marginal Probability
Probability theory Naïve Bayes classification Random variable: a variable whose possible values are numerical outcomes of a random phenomenon. s: A person s height, the outcome of a coin toss Distinguish
More informationMultivariate Bayesian Linear Regression MLAI Lecture 11
Multivariate Bayesian Linear Regression MLAI Lecture 11 Neil D. Lawrence Department of Computer Science Sheffield University 21st October 2012 Outline Univariate Bayesian Linear Regression Multivariate
More information2.4 Conditional Probability
2.4 Conditional Probability The probabilities assigned to various events depend on what is known about the experimental situation when the assignment is made. Example: Suppose a pair of dice is tossed.
More informationCOURSE INTRODUCTION. J. Elder CSE 6390/PSYC 6225 Computational Modeling of Visual Perception
COURSE INTRODUCTION COMPUTATIONAL MODELING OF VISUAL PERCEPTION 2 The goal of this course is to provide a framework and computational tools for modeling visual inference, motivated by interesting examples
More informationSTAT 418: Probability and Stochastic Processes
STAT 418: Probability and Stochastic Processes Spring 2016; Homework Assignments Latest updated on April 29, 2016 HW1 (Due on Jan. 21) Chapter 1 Problems 1, 8, 9, 10, 11, 18, 19, 26, 28, 30 Theoretical
More informationBe able to define the following terms and answer basic questions about them:
CS440/ECE448 Section Q Fall 2017 Final Review Be able to define the following terms and answer basic questions about them: Probability o Random variables, axioms of probability o Joint, marginal, conditional
More informationAn event described by a single characteristic e.g., A day in January from all days in 2012
Events Each possible outcome of a variable is an event. Simple event An event described by a single characteristic e.g., A day in January from all days in 2012 Joint event An event described by two or
More informationChapter 3: Maximum-Likelihood & Bayesian Parameter Estimation (part 1)
HW 1 due today Parameter Estimation Biometrics CSE 190 Lecture 7 Today s lecture was on the blackboard. These slides are an alternative presentation of the material. CSE190, Winter10 CSE190, Winter10 Chapter
More informationCurve Fitting Re-visited, Bishop1.2.5
Curve Fitting Re-visited, Bishop1.2.5 Maximum Likelihood Bishop 1.2.5 Model Likelihood differentiation p(t x, w, β) = Maximum Likelihood N N ( t n y(x n, w), β 1). (1.61) n=1 As we did in the case of the
More informationProbability and (Bayesian) Data Analysis
Department of Statistics The University of Auckland https://www.stat.auckland.ac.nz/ brewer/ Where to get everything To get all of the material (slides, code, exercises): git clone --recursive https://github.com/eggplantbren/madrid
More informationProbability the chance that an uncertain event will occur (always between 0 and 1)
Quantitative Methods 2013 1 Probability as a Numerical Measure of the Likelihood of Occurrence Probability the chance that an uncertain event will occur (always between 0 and 1) Increasing Likelihood of
More informationNotes on Machine Learning for and
Notes on Machine Learning for 16.410 and 16.413 (Notes adapted from Tom Mitchell and Andrew Moore.) Choosing Hypotheses Generally want the most probable hypothesis given the training data Maximum a posteriori
More informationarxiv:astro-ph/ v1 14 Sep 2005
For publication in Bayesian Inference and Maximum Entropy Methods, San Jose 25, K. H. Knuth, A. E. Abbas, R. D. Morris, J. P. Castle (eds.), AIP Conference Proceeding A Bayesian Analysis of Extrasolar
More informationExpectation maximization
Expectation maximization Subhransu Maji CMSCI 689: Machine Learning 14 April 2015 Motivation Suppose you are building a naive Bayes spam classifier. After your are done your boss tells you that there is
More informationMathematical Formulation of Our Example
Mathematical Formulation of Our Example We define two binary random variables: open and, where is light on or light off. Our question is: What is? Computer Vision 1 Combining Evidence Suppose our robot
More informationConditional Probability, Total Probability Theorem and Bayes Rule
Conditional robability, Total robability Theorem and Bayes Rule Berlin Chen Department of Computer Science & Information Engineering National Taiwan Normal University Reference: - D.. Bertsekas, J. N.
More informationProbability, Conditional Probability and Bayes Rule IE231 - Lecture Notes 3 Mar 6, 2018
Probability, Conditional Probability and Bayes Rule IE31 - Lecture Notes 3 Mar 6, 018 #Introduction Let s recall some probability concepts. Probability is the quantification of uncertainty. For instance
More informationConditional probability
CHAPTER 4 Conditional probability 4.1. Introduction Suppose there are 200 men, of which 100 are smokers, and 100 women, of which 20 are smokers. What is the probability that a person chosen at random will
More informationMachine Learning
Machine Learning 10-601 Tom M. Mitchell Machine Learning Department Carnegie Mellon University August 30, 2017 Today: Decision trees Overfitting The Big Picture Coming soon Probabilistic learning MLE,
More informationOutline. Binomial, Multinomial, Normal, Beta, Dirichlet. Posterior mean, MAP, credible interval, posterior distribution
Outline A short review on Bayesian analysis. Binomial, Multinomial, Normal, Beta, Dirichlet Posterior mean, MAP, credible interval, posterior distribution Gibbs sampling Revisit the Gaussian mixture model
More informationComputer Vision Group Prof. Daniel Cremers. 3. Regression
Prof. Daniel Cremers 3. Regression Categories of Learning (Rep.) Learnin g Unsupervise d Learning Clustering, density estimation Supervised Learning learning from a training data set, inference on the
More informationParametric Models. Dr. Shuang LIANG. School of Software Engineering TongJi University Fall, 2012
Parametric Models Dr. Shuang LIANG School of Software Engineering TongJi University Fall, 2012 Today s Topics Maximum Likelihood Estimation Bayesian Density Estimation Today s Topics Maximum Likelihood
More informationLecture 01: Introduction
Lecture 01: Introduction Dipankar Bandyopadhyay, Ph.D. BMTRY 711: Analysis of Categorical Data Spring 2011 Division of Biostatistics and Epidemiology Medical University of South Carolina Lecture 01: Introduction
More informationwe need to describe how many cookies the first person gets. There are 6 choices (0, 1,... 5). So the answer is 6.
() (a) How many ways are there to divide 5 different cakes and 5 identical cookies between people so that the first person gets exactly cakes. (b) How many ways are there to divide 5 different cakes and
More informationOverfitting, Bias / Variance Analysis
Overfitting, Bias / Variance Analysis Professor Ameet Talwalkar Professor Ameet Talwalkar CS260 Machine Learning Algorithms February 8, 207 / 40 Outline Administration 2 Review of last lecture 3 Basic
More informationan introduction to bayesian inference
with an application to network analysis http://jakehofman.com january 13, 2010 motivation would like models that: provide predictive and explanatory power are complex enough to describe observed phenomena
More informationIntroduction to Statistical Methods for High Energy Physics
Introduction to Statistical Methods for High Energy Physics 2011 CERN Summer Student Lectures Glen Cowan Physics Department Royal Holloway, University of London g.cowan@rhul.ac.uk www.pp.rhul.ac.uk/~cowan
More informationDetermining Probabilities. Product Rule for Ordered Pairs/k-Tuples:
Determining Probabilities Product Rule for Ordered Pairs/k-Tuples: Determining Probabilities Product Rule for Ordered Pairs/k-Tuples: Proposition If the first element of object of an ordered pair can be
More information