Monte Carlo Simulation I
|
|
- Sara Cook
- 5 years ago
- Views:
Transcription
1 Monte Carlo Simulation I Anders Ringgaard Kristensen
2 What is simulation? Simulation is an attempt to model a real world system in order to: Obtain a better understanding of the system (including interactions) Fysiological models Herd models Study the effects of various (complex) decision strategies Herd models Slide 2
3 Different kinds of simulation Randomness (cf. Chapter 14.3 in textbook) Deterministic forget it in herd management! All calculations based on average values Same input Same output System comprehension Differential equations, physiology (NorFor). Probabilistic models All calculations based on distributions Same input Same output System comprehension, decision strategies Markov chains (earlier this course) Stochastic (Monte Carlo) models All random events are simulated by random number generation Same input Different output i.e. we need many replications System comprehension, decision strategies Herd constraints, complexity Slide 3
4 Different kinds of simulation Hierarchy (levels) Mechanistic: A system is modeled by its elements (sub-systems) A herd is modeled by its individual animals and their interactions. Empirical Only one level modeled Output direcly modeled from output Time Dynamic Static Slide 4
5 Our main example: SimFlock A stochastic, mechanistic, dynamic simulation model of a scavenging chicken production system. Smallholder production most often women. A very direct modeling of the real system Slide 5
6 SimFlock: An object oriented model Breeding animals Hens & Cocks Eggs Chicks Growers Infertile Dead Household consumption Pullets Cockerels Slide 6 Market
7 User interface visible objects All birds and eggs present in the flock shown. States of the birds can be investigated Demo
8 Other (Danish) herd simulation models SimHerd (dairy herds case example later) Jan Tind Sørensen Søren Østergaard Anne Braad Kudahl Jehan Ettema The Dina Pig simulation model Erik Jørgensen Slide 8
9 The TV Show once again Earlier we solved the decision problem by use of Bayesian networks and/or decision graphs. Are there any other methods we could apply? Experiment many replications needed Simulation create a simulation model
10 How to model the problem Identify the variables: True placement, True {1, 2, 3} First choice, Choice1 {1, 2, 3} Door opened, Opened {1, 2, 3} Second choice, Choice 2 {Keep, Change} Reward, Gain {0, 1000} Define a decision strategy (2 options): Choice2 = Keep Choice2 = Change Slide 10
11 What is needed for simulation Coin: Probabilities A random number generator, e.g. A coin A dice A computer 1 0,8 0,6 0,4 0, Dice: Probabilities Computer: Density 1 1,2 0,8 0,6 0,4 0,2 1 0,8 0,6 0,4 0, ,2 0,4 0,6 0,8 1
12 Simulation procedure: Use dice Host must place the reward behind an arbitrarily selected door: Roll the dice: 1 or 2: Door 1 3 or 4: Door 2 5 or 6: Door 3 Value of variable True determined as 1, 2 or 3! Participant must choose a door at random Roll the dice: 1 or 2: Door 1 3 or 4: Door 2 5 or 6: Door 3 Value of variable Choice 1 determined as 1, 2 or 3! Slide 12
13 Simulation procedure II Check whether True = Choice 1 If yes (two options): Roll the dice (or toss a coin) 1, 2 or 3: Open the lowest door number i True 4, 5 or 6: Open the highest door number i True If no (only 1 option): Open door i True and i Choice 1 Value of Opened determined Slide 13
14 Simulation procedure III Value of Choice 2 determined in accordance with the decision strategy Define new variable Final guess {1, 2, 3} If Choice 2 = Keep: Final guess = Choice 1 If Choice 2 = Change: Final guess = i, where i Choice 1 and i Opened Slide 14
15 Simulation procedure IV Check whether Final guess = True If yes: Gain = 1000 If no: Gain = 0 Simulation completed! Slide 15
16 Evaluation of strategies Define the strategy as Choice 2 = Keep Repeat the simulation many times (e.g. 1000) and calculate the average gain under the strategy. Define the strategy as Choice 2 = Change Repeat the simulation many times (e.g. 1000) and calculate the average gain under the strategy. Compare the average gain under the two strategies and select the best. Slide 16
17 Simulation procedure: Use computer Exactly as before, except: Instead of rolling the dice, we let the computer draw a random number r. Converting to variable value for True is done as follows: If r < : Door 1 If < r < : Door 2 If r > : Door 3 (Similar for other variable values) Let s try! Slide 17
18 Formalizing the simulation We have (refer to Chapter 14.2 of the textbook) A decision rule Θ: Very simple, just Keep or Change A set of parameters Φ = (Φ 0, Φ s. ) State of nature Φ 0 we have 4, p t1, p t2, p c1, p c2 : p t1 = P(True = 1), p t2 = P(True = 2), but P(True = 3) = 1 -p t1 -p t2 p c1 = P(Ch1= 1), p c2 = P(Ch1 = 2), but P(Ch1= 3) = 1 -p c1 -p c2 State variables, describing the system during simulation: Φ s. : True, Choice 1, Opened, Gain An output variable Ω: Gain. Notice that Ω Φ s. Model input is (Φ 0, Θ). Slide 18
19 Purpose of simulation The purpose of simulation usually is to calculate the expected utility E(U(Θ, Φ)) under a certain decision rule, Θ. In the example it is the expected gain under a specified decision rule (Keep or Change) This may be regarded as solving the numerical integration problem: Slide 19
20 The integral expression Horrible that s why we choose to build a simulation model instead! Let us take a look at the elements: Φ = (p t1, p t2, p c1, p c2, True,Choice 1, Opened, Gain) is a complete set of parameters. A value could for instance be φ = (1/3, 1/3, 1/3, 1/3, 2, 1, 3, 1000). Notice that we can easily specify the utility for a given outcome: U(Θ, Φ = φ) = Gain If we wanted to solve the integral directly we should furthermore be able to calculate the combined probability p(φ = φ) for any legal outcome φ of Φ. Almost impossible! Slide 20
21 The numerical approximation Simulation: Generate (through simulation) e.g. k outcomes φ 1, φ 2,, φ k under a chosen decision rule Θ: φ 1 = (1/3, 1/3, 1/3, 1/3, 2, 1, 3, 1000), U(Θ, φ 1 ) = 1000 φ 2 = (1/3, 1/3, 1/3, 1/3, 3, 3, 1, 0), U(Θ, φ 2 ) = 0 φ k = (1/3, 1/3, 1/3, 1/3, 1, 3, 2, 1000), U(Θ, φ k ) = 1000 Approximate the integral with the average utility: Slide 21
22 State of nature, I In the example the state of nature has been regarded as fixed and known Φ 0 = (1/3, 1/3, 1/3, 1/3). Assume that the host has a favorite door, e.g. door 3. He places the reward behind Door 3 with probability 0.8 and behind each of the others with probability 0.1. State of nature under those circumstances would be Φ 0 = (0.1, 0.1, 1/3, 1/3) if the participant doesn t have a favorite door. Does it change anything? Let s try! Slide 22
23 State of nature, II Assume further that also the participant has a favorite door, e.g. door 3. He also selects (first choice) Door 3 with probability 0.8 and each of the others with probability 0.1. State of nature under those circumstances would be Φ 0 = (0.1, 0.1, 0.1, 0.1) when both have Door 3 as their favorite door. Does it change anything? Let s try! Slide 23
24 State of nature, III In general, the optimal decision rule as well as the expected result depend on the state of nature. What is the state of nature in a livestock simulation model: Average growth rate Herd mortality rate Average milk yield Is the true state of nature known (with certainty)? Does it matter? Slide 24
25 State of nature, livestock models In livestock models we never know the true state of nature. We need to represent the uncertainty of the state of nature. We typically have some ideas a belief in the true values. The belief may be represented as a statistical distribution. Slide 25
26 Simple example, I Slide 26
27 Simple example, II Evaluation of expected value of production Deterministic: Stochastic, known state of nature Simulate 100 replications with µ = 10 Average: 435,27; Std. Dev Stochastic, unknown state of nature: Draw 100 random values of µ from N(10, 1) For each µ, simulate 100 replications Advanced Quantitative Average: Methods in Herd Management 430, Std. Dev Slide 27
28 Simple example, III 0,4 Frequency 0,3 0,2 0, Value of production Slide 28 Know n state of nature Unknow n state of nature
29 Representation of uncertainty of SON Essential, but Most simulation models ignore it Since the parameters Φ = (Φ 0, Φ s. ) consist of state of nature parameters and state variables (changing during simulation) we may split the integration from before: Slide 29
30 Interpretation Innermost integral: The expected utility for a given state of nature (Φ 0 = φ 0 ). Outermost integral: The expected utility across states of nature. Slide 30
31 Simulation procedure Select a number of alternative decision rules Θ to be tested. Draw n states of nature φ 01, φ 02,, φ 0 n from the underlying distribution of state of nature p 0 (Φ 0 =φ 0 ) For each decision rule: For each state of nature, i: Run the simulation model m times. Calculate the average result U i over the m simulation results (numerical evaluation of the innermost integral). Calculate the average value of U 1, U 2,, U n (numerical evaluation of the outermost integral). Select the best performing decision rule. Slide 31
32 Main problem It is difficult to specify the distribution of the state of nature. For a systematic description of the approach used in the SimFlock model, reference is made to Kristensen & Pedersen (2003) link at the homepage. Basic principle: Each parameter of the state of nature is specified through a distribution instead of a value. Such a distribution is called a hyper distribution. The parameters of a hyper distribution are called hyper parameters. Estimated from production data from 30 flocks in Zimbabwe. Easy, if parameters are independent Difficult if they interact Wednesday we shall take a closer look! Slide 32
10/3/2018. Our main example: SimFlock. Breeding animals Hens & Cocks
What is simulation? Monte Carlo Simulation I Anders Ringgaard Kristensen Simulation is an attempt to model a real world system in order to: Obtain a better understanding of the system (including interactions)
More informationIntroduction to Bayesian Networks
Introduction to Bayesian Networks Anders Ringgaard Kristensen Slide 1 Outline Causal networks Bayesian Networks Evidence Conditional Independence and d-separation Compilation The moral graph The triangulated
More informationOutline. A quiz
Introduction to Bayesian Networks Anders Ringgaard Kristensen Outline Causal networks Bayesian Networks Evidence Conditional Independence and d-separation Compilation The moral graph The triangulated graph
More informationAdvanced Herd Management Probabilities and distributions
Advanced Herd Management Probabilities and distributions Anders Ringgaard Kristensen Slide 1 Outline Probabilities Conditional probabilities Bayes theorem Distributions Discrete Continuous Distribution
More informationAdvanced topics from statistics
Advanced topics from statistics Anders Ringgaard Kristensen Advanced Herd Management Slide 1 Outline Covariance and correlation Random vectors and multivariate distributions The multinomial distribution
More informationElementary Discrete Probability
Elementary Discrete Probability MATH 472 Financial Mathematics J Robert Buchanan 2018 Objectives In this lesson we will learn: the terminology of elementary probability, elementary rules of probability,
More informationProbability Rules. MATH 130, Elements of Statistics I. J. Robert Buchanan. Fall Department of Mathematics
Probability Rules MATH 130, Elements of Statistics I J. Robert Buchanan Department of Mathematics Fall 2018 Introduction Probability is a measure of the likelihood of the occurrence of a certain behavior
More informationStatistics Introduction to Probability
Statistics 110 - Introduction to Probability Mark E. Irwin Department of Statistics Harvard University Summer Term Monday, June 26, 2006 - Wednesday, August 16, 2006 Copyright 2006 by Mark E. Irwin Personnel
More informationIntroduction to Artificial Intelligence (AI)
Introduction to Artificial Intelligence (AI) Computer Science cpsc502, Lecture 9 Oct, 11, 2011 Slide credit Approx. Inference : S. Thrun, P, Norvig, D. Klein CPSC 502, Lecture 9 Slide 1 Today Oct 11 Bayesian
More informationLecture 15. DATA 8 Spring Sampling. Slides created by John DeNero and Ani Adhikari
DATA 8 Spring 2018 Lecture 15 Sampling Slides created by John DeNero (denero@berkeley.edu) and Ani Adhikari (adhikari@berkeley.edu) Announcements Probability Basics Lowest value: 0 Chance of event that
More informationBayesian networks II. Model building. Anders Ringgaard Kristensen
Bayesian networks II. Model building Anders Ringgaard Kristensen Outline Determining the graphical structure Milk test Mastitis diagnosis Pregnancy Determining the conditional probabilities Modeling methods
More informationBayesian inference. Fredrik Ronquist and Peter Beerli. October 3, 2007
Bayesian inference Fredrik Ronquist and Peter Beerli October 3, 2007 1 Introduction The last few decades has seen a growing interest in Bayesian inference, an alternative approach to statistical inference.
More informationA(VeryBrief)IntroductiontoIdealPoint Estimates
A(VeryBrief)IntroductiontoIdealPoint Estimates 1 Motivation 2 3 4 5 Motivation A cab was involved in a hit-and run accident at night. Two cab companies (Green and Blue) operate the city. 85% of the cabs
More informationDepartment of Large Animal Sciences. Outline. Slide 2. Department of Large Animal Sciences. Slide 4. Department of Large Animal Sciences
Outline Advanced topics from statistics Anders Ringgaard Kristensen Covariance and correlation Random vectors and multivariate distributions The multinomial distribution The multivariate normal distribution
More informationDiscrete Mathematics and Probability Theory Spring 2014 Anant Sahai Note 10
EECS 70 Discrete Mathematics and Probability Theory Spring 2014 Anant Sahai Note 10 Introduction to Basic Discrete Probability In the last note we considered the probabilistic experiment where we flipped
More informationApproximate Inference
Approximate Inference Simulation has a name: sampling Sampling is a hot topic in machine learning, and it s really simple Basic idea: Draw N samples from a sampling distribution S Compute an approximate
More informationTutorial on Probabilistic Programming with PyMC3
185.A83 Machine Learning for Health Informatics 2017S, VU, 2.0 h, 3.0 ECTS Tutorial 02-04.04.2017 Tutorial on Probabilistic Programming with PyMC3 florian.endel@tuwien.ac.at http://hci-kdd.org/machine-learning-for-health-informatics-course
More informationIntroduction to Artificial Intelligence (AI)
Introduction to Artificial Intelligence (AI) Computer Science cpsc502, Lecture 10 Oct, 13, 2011 CPSC 502, Lecture 10 Slide 1 Today Oct 13 Inference in HMMs More on Robot Localization CPSC 502, Lecture
More informationStatistics Introduction to Probability
Statistics 110 - Introduction to Probability Mark E. Irwin Department of Statistics Harvard University Summer Term Monday, June 28, 2004 - Wednesday, August 18, 2004 Personnel Instructor: Mark Irwin Office:
More informationDiscrete Random Variables
Discrete Random Variables An Undergraduate Introduction to Financial Mathematics J. Robert Buchanan 2014 Introduction The markets can be thought of as a complex interaction of a large number of random
More informationMATH MW Elementary Probability Course Notes Part I: Models and Counting
MATH 2030 3.00MW Elementary Probability Course Notes Part I: Models and Counting Tom Salisbury salt@yorku.ca York University Winter 2010 Introduction [Jan 5] Probability: the mathematics used for Statistics
More informationDynamic Risk Measures and Nonlinear Expectations with Markov Chain noise
Dynamic Risk Measures and Nonlinear Expectations with Markov Chain noise Robert J. Elliott 1 Samuel N. Cohen 2 1 Department of Commerce, University of South Australia 2 Mathematical Insitute, University
More informationReasoning with Uncertainty
Reasoning with Uncertainty Representing Uncertainty Manfred Huber 2005 1 Reasoning with Uncertainty The goal of reasoning is usually to: Determine the state of the world Determine what actions to take
More informationCS 4649/7649 Robot Intelligence: Planning
CS 4649/7649 Robot Intelligence: Planning Probability Primer Sungmoon Joo School of Interactive Computing College of Computing Georgia Institute of Technology S. Joo (sungmoon.joo@cc.gatech.edu) 1 *Slides
More informationIntroduction to Mobile Robotics Probabilistic Robotics
Introduction to Mobile Robotics Probabilistic Robotics Wolfram Burgard 1 Probabilistic Robotics Key idea: Explicit representation of uncertainty (using the calculus of probability theory) Perception Action
More informationLecture Slides - Part 1
Lecture Slides - Part 1 Bengt Holmstrom MIT February 2, 2016. Bengt Holmstrom (MIT) Lecture Slides - Part 1 February 2, 2016. 1 / 36 Going to raise the level a little because 14.281 is now taught by Juuso
More informationProbability and Time: Hidden Markov Models (HMMs)
Probability and Time: Hidden Markov Models (HMMs) Computer Science cpsc322, Lecture 32 (Textbook Chpt 6.5.2) Nov, 25, 2013 CPSC 322, Lecture 32 Slide 1 Lecture Overview Recap Markov Models Markov Chain
More informationThe Theory behind PageRank
The Theory behind PageRank Mauro Sozio Telecom ParisTech May 21, 2014 Mauro Sozio (LTCI TPT) The Theory behind PageRank May 21, 2014 1 / 19 A Crash Course on Discrete Probability Events and Probability
More informationORF 245 Fundamentals of Statistics Chapter 1 Probability
ORF 245 Fundamentals of Statistics Chapter 1 Probability Robert Vanderbei Sept 2014 Slides last edited on September 19, 2014 http://www.princeton.edu/ rvdb Course Info Prereqs: Textbook: Three semesters
More informationReasoning Under Uncertainty: Introduction to Probability
Reasoning Under Uncertainty: Introduction to Probability CPSC 322 Uncertainty 1 Textbook 6.1 Reasoning Under Uncertainty: Introduction to Probability CPSC 322 Uncertainty 1, Slide 1 Lecture Overview 1
More informationBayesian Inference and MCMC
Bayesian Inference and MCMC Aryan Arbabi Partly based on MCMC slides from CSC412 Fall 2018 1 / 18 Bayesian Inference - Motivation Consider we have a data set D = {x 1,..., x n }. E.g each x i can be the
More informationMarkov localization uses an explicit, discrete representation for the probability of all position in the state space.
Markov Kalman Filter Localization Markov localization localization starting from any unknown position recovers from ambiguous situation. However, to update the probability of all positions within the whole
More informationCS 188: Artificial Intelligence Spring Today
CS 188: Artificial Intelligence Spring 2006 Lecture 9: Naïve Bayes 2/14/2006 Dan Klein UC Berkeley Many slides from either Stuart Russell or Andrew Moore Bayes rule Today Expectations and utilities Naïve
More informationLecture 3. January 7, () Lecture 3 January 7, / 35
Lecture 3 January 7, 2013 () Lecture 3 January 7, 2013 1 / 35 Outline This week s lecture: Fast review of last week s lecture: Conditional probability. Partition, Partition theorem. Bayes theorem and its
More informationMath 105 Course Outline
Math 105 Course Outline Week 9 Overview This week we give a very brief introduction to random variables and probability theory. Most observable phenomena have at least some element of randomness associated
More informationIntroductory Econometrics. Review of statistics (Part II: Inference)
Introductory Econometrics Review of statistics (Part II: Inference) Jun Ma School of Economics Renmin University of China October 1, 2018 1/16 Null and alternative hypotheses Usually, we have two competing
More informationBayesian Methods for Machine Learning
Bayesian Methods for Machine Learning CS 584: Big Data Analytics Material adapted from Radford Neal s tutorial (http://ftp.cs.utoronto.ca/pub/radford/bayes-tut.pdf), Zoubin Ghahramni (http://hunch.net/~coms-4771/zoubin_ghahramani_bayesian_learning.pdf),
More informationQuiz 1. Name: Instructions: Closed book, notes, and no electronic devices.
Quiz 1. Name: Instructions: Closed book, notes, and no electronic devices. 1. What is the difference between a deterministic model and a probabilistic model? (Two or three sentences only). 2. What is the
More informationBayesian Inference. Anders Gorm Pedersen. Molecular Evolution Group Center for Biological Sequence Analysis Technical University of Denmark (DTU)
Bayesian Inference Anders Gorm Pedersen Molecular Evolution Group Center for Biological Sequence Analysis Technical University of Denmark (DTU) Background: Conditional probability A P (B A) = A,B P (A,
More informationProbabilistic Machine Learning
Probabilistic Machine Learning Bayesian Nets, MCMC, and more Marek Petrik 4/18/2017 Based on: P. Murphy, K. (2012). Machine Learning: A Probabilistic Perspective. Chapter 10. Conditional Independence Independent
More informationChapter 7 Probability Basics
Making Hard Decisions Chapter 7 Probability Basics Slide 1 of 62 Introduction A A,, 1 An Let be an event with possible outcomes: 1 A = Flipping a coin A = {Heads} A 2 = Ω {Tails} The total event (or sample
More informationSUPPLEMENTARY TABLES FOR: THE LIKELIHOOD RATIO TEST FOR COINTEGRATION RANKS IN THE I(2) MODEL
SUPPLEMENTARY TABLES FOR: THE LIKELIHOOD RATIO TEST FOR COINTEGRATION RANKS IN THE I(2) MODEL Heino Bohn Nielsen and Anders Rahbek Department of Economics University of Copenhagen. heino.bohn.nielsen@econ.ku.dk
More informationIntroduction to Bayesian Inference
Introduction to Bayesian Inference p. 1/2 Introduction to Bayesian Inference September 15th, 2010 Reading: Hoff Chapter 1-2 Introduction to Bayesian Inference p. 2/2 Probability: Measurement of Uncertainty
More informationOn the errors introduced by the naive Bayes independence assumption
On the errors introduced by the naive Bayes independence assumption Author Matthijs de Wachter 3671100 Utrecht University Master Thesis Artificial Intelligence Supervisor Dr. Silja Renooij Department of
More information6 Reinforcement Learning
6 Reinforcement Learning As discussed above, a basic form of supervised learning is function approximation, relating input vectors to output vectors, or, more generally, finding density functions p(y,
More informationLecture 1. ABC of Probability
Math 408 - Mathematical Statistics Lecture 1. ABC of Probability January 16, 2013 Konstantin Zuev (USC) Math 408, Lecture 1 January 16, 2013 1 / 9 Agenda Sample Spaces Realizations, Events Axioms of Probability
More information1 Ways to Describe a Stochastic Process
purdue university cs 59000-nmc networks & matrix computations LECTURE NOTES David F. Gleich September 22, 2011 Scribe Notes: Debbie Perouli 1 Ways to Describe a Stochastic Process We will use the biased
More informationAnnouncements. CS 188: Artificial Intelligence Fall Adversarial Games. Computing Minimax Values. Evaluation Functions. Recap: Resource Limits
CS 188: Artificial Intelligence Fall 2009 Lecture 7: Expectimax Search 9/17/2008 Announcements Written 1: Search and CSPs is up Project 2: Multi-agent Search is up Want a partner? Come to the front after
More informationECE295, Data Assimila0on and Inverse Problems, Spring 2015
ECE295, Data Assimila0on and Inverse Problems, Spring 2015 1 April, Intro; Linear discrete Inverse problems (Aster Ch 1 and 2) Slides 8 April, SVD (Aster ch 2 and 3) Slides 15 April, RegularizaFon (ch
More informationANALYTIC COMPARISON. Pearl and Rubin CAUSAL FRAMEWORKS
ANALYTIC COMPARISON of Pearl and Rubin CAUSAL FRAMEWORKS Content Page Part I. General Considerations Chapter 1. What is the question? 16 Introduction 16 1. Randomization 17 1.1 An Example of Randomization
More informationDecision Theory: Q-Learning
Decision Theory: Q-Learning CPSC 322 Decision Theory 5 Textbook 12.5 Decision Theory: Q-Learning CPSC 322 Decision Theory 5, Slide 1 Lecture Overview 1 Recap 2 Asynchronous Value Iteration 3 Q-Learning
More informationLecture 3: Policy Evaluation Without Knowing How the World Works / Model Free Policy Evaluation
Lecture 3: Policy Evaluation Without Knowing How the World Works / Model Free Policy Evaluation CS234: RL Emma Brunskill Winter 2018 Material builds on structure from David SIlver s Lecture 4: Model-Free
More informationFormalizing Probability. Choosing the Sample Space. Probability Measures
Formalizing Probability Choosing the Sample Space What do we assign probability to? Intuitively, we assign them to possible events (things that might happen, outcomes of an experiment) Formally, we take
More informationStatistical Methods in Particle Physics Lecture 1: Bayesian methods
Statistical Methods in Particle Physics Lecture 1: Bayesian methods SUSSP65 St Andrews 16 29 August 2009 Glen Cowan Physics Department Royal Holloway, University of London g.cowan@rhul.ac.uk www.pp.rhul.ac.uk/~cowan
More informationProbability Basics. Part 3: Types of Probability. INFO-1301, Quantitative Reasoning 1 University of Colorado Boulder
Probability Basics Part 3: Types of Probability INFO-1301, Quantitative Reasoning 1 University of Colorado Boulder September 30, 2016 Prof. Michael Paul Prof. William Aspray Example A large government
More informationA Tutorial on Learning with Bayesian Networks
A utorial on Learning with Bayesian Networks David Heckerman Presented by: Krishna V Chengavalli April 21 2003 Outline Introduction Different Approaches Bayesian Networks Learning Probabilities and Structure
More informationStats Probability Theory
Stats 241.3 Probability Theory Instructor: Office: W.H.Laverty 235 McLean Hall Phone: 966-6096 Lectures: Evaluation: M T W Th F 1:30pm - 2:50pm Thorv 105 Lab: T W Th 3:00-3:50 Thorv 105 Assignments, Labs,
More informationUncertainty. Michael Peters December 27, 2013
Uncertainty Michael Peters December 27, 20 Lotteries In many problems in economics, people are forced to make decisions without knowing exactly what the consequences will be. For example, when you buy
More information10. Exchangeability and hierarchical models Objective. Recommended reading
10. Exchangeability and hierarchical models Objective Introduce exchangeability and its relation to Bayesian hierarchical models. Show how to fit such models using fully and empirical Bayesian methods.
More informationSTT When trying to evaluate the likelihood of random events we are using following wording.
Introduction to Chapter 2. Probability. When trying to evaluate the likelihood of random events we are using following wording. Provide your own corresponding examples. Subjective probability. An individual
More informationSatisfaction Equilibrium: Achieving Cooperation in Incomplete Information Games
Satisfaction Equilibrium: Achieving Cooperation in Incomplete Information Games Stéphane Ross and Brahim Chaib-draa Department of Computer Science and Software Engineering Laval University, Québec (Qc),
More informationChapter 1 (Basic Probability)
Chapter 1 (Basic Probability) What is probability? Consider the following experiments: 1. Count the number of arrival requests to a web server in a day. 2. Determine the execution time of a program. 3.
More informationOutline. CSE 573: Artificial Intelligence Autumn Agent. Partial Observability. Markov Decision Process (MDP) 10/31/2012
CSE 573: Artificial Intelligence Autumn 2012 Reasoning about Uncertainty & Hidden Markov Models Daniel Weld Many slides adapted from Dan Klein, Stuart Russell, Andrew Moore & Luke Zettlemoyer 1 Outline
More informationInformation Retrieval and Web Search Engines
Information Retrieval and Web Search Engines Lecture 4: Probabilistic Retrieval Models April 29, 2010 Wolf-Tilo Balke and Joachim Selke Institut für Informationssysteme Technische Universität Braunschweig
More informationTutorial 2: Probability
1 Tutorial 2: Probability 1. Coin Tosses Recall that the probability of a specified outcome or outcome(s) is given by prob. of success = number of successful outcomes total number of possible outcomes.
More informationProbabilistic Robotics
Probabilistic Robotics Overview of probability, Representing uncertainty Propagation of uncertainty, Bayes Rule Application to Localization and Mapping Slides from Autonomous Robots (Siegwart and Nourbaksh),
More informationPreliminary Statistics Lecture 2: Probability Theory (Outline) prelimsoas.webs.com
1 School of Oriental and African Studies September 2015 Department of Economics Preliminary Statistics Lecture 2: Probability Theory (Outline) prelimsoas.webs.com Gujarati D. Basic Econometrics, Appendix
More informationRandomized Algorithms
Randomized Algorithms Prof. Tapio Elomaa tapio.elomaa@tut.fi Course Basics A new 4 credit unit course Part of Theoretical Computer Science courses at the Department of Mathematics There will be 4 hours
More informationMarkov Chains. Chapter 16. Markov Chains - 1
Markov Chains Chapter 16 Markov Chains - 1 Why Study Markov Chains? Decision Analysis focuses on decision making in the face of uncertainty about one future event. However, many decisions need to consider
More informationLecture 3: Probabilistic Retrieval Models
Probabilistic Retrieval Models Information Retrieval and Web Search Engines Lecture 3: Probabilistic Retrieval Models November 5 th, 2013 Wolf-Tilo Balke and Kinda El Maarry Institut für Informationssysteme
More informationLecture 5. G. Cowan Lectures on Statistical Data Analysis Lecture 5 page 1
Lecture 5 1 Probability (90 min.) Definition, Bayes theorem, probability densities and their properties, catalogue of pdfs, Monte Carlo 2 Statistical tests (90 min.) general concepts, test statistics,
More informationEvaluation for Pacman. CS 188: Artificial Intelligence Fall Iterative Deepening. α-β Pruning Example. α-β Pruning Pseudocode.
CS 188: Artificial Intelligence Fall 2008 Evaluation for Pacman Lecture 7: Expectimax Search 9/18/2008 [DEMO: thrashing, smart ghosts] Dan Klein UC Berkeley Many slides over the course adapted from either
More informationCS 188: Artificial Intelligence Fall 2008
CS 188: Artificial Intelligence Fall 2008 Lecture 7: Expectimax Search 9/18/2008 Dan Klein UC Berkeley Many slides over the course adapted from either Stuart Russell or Andrew Moore 1 1 Evaluation for
More informationLogic: Intro & Propositional Definite Clause Logic
Logic: Intro & Propositional Definite Clause Logic Alan Mackworth UBC CS 322 Logic 1 February 27, 2013 P & M extbook 5.1 Lecture Overview Recap: CSP planning Intro to Logic Propositional Definite Clause
More informationWhere are we in CS 440?
Where are we in CS 440? Now leaving: sequential deterministic reasoning Entering: probabilistic reasoning and machine learning robability: Review of main concepts Chapter 3 Making decisions under uncertainty
More informationDecision Theory: Markov Decision Processes
Decision Theory: Markov Decision Processes CPSC 322 Lecture 33 March 31, 2006 Textbook 12.5 Decision Theory: Markov Decision Processes CPSC 322 Lecture 33, Slide 1 Lecture Overview Recap Rewards and Policies
More informationCS 188: Artificial Intelligence Fall Announcements
CS 188: Artificial Intelligence Fall 2009 Lecture 7: Expectimax Search 9/17/2008 Dan Klein UC Berkeley Many slides over the course adapted from either Stuart Russell or Andrew Moore 1 Announcements Written
More information11. Probability Sample Spaces and Probability
11. Probability 11.1 Sample Spaces and Probability 1 Objectives A. Find the probability of an event. B. Find the empirical probability of an event. 2 Theoretical Probabilities 3 Example A fair coin is
More informationProbabilistic Reasoning
Course 16 :198 :520 : Introduction To Artificial Intelligence Lecture 7 Probabilistic Reasoning Abdeslam Boularias Monday, September 28, 2015 1 / 17 Outline We show how to reason and act under uncertainty.
More informationPROBABILITY THEORY. Prof. S. J. Soni. Assistant Professor Computer Engg. Department SPCE, Visnagar
PROBABILITY THEORY By Prof. S. J. Soni Assistant Professor Computer Engg. Department SPCE, Visnagar Introduction Signals whose values at any instant t are determined by their analytical or graphical description
More informationOther Noninformative Priors
Other Noninformative Priors Other methods for noninformative priors include Bernardo s reference prior, which seeks a prior that will maximize the discrepancy between the prior and the posterior and minimize
More informationRandomized Complexity Classes; RP
Randomized Complexity Classes; RP Let N be a polynomial-time precise NTM that runs in time p(n) and has 2 nondeterministic choices at each step. N is a polynomial Monte Carlo Turing machine for a language
More informationLearning the hyper-parameters. Luca Martino
Learning the hyper-parameters Luca Martino 2017 2017 1 / 28 Parameters and hyper-parameters 1. All the described methods depend on some choice of hyper-parameters... 2. For instance, do you recall λ (bandwidth
More informationMAE 493G, CpE 493M, Mobile Robotics. 6. Basic Probability
MAE 493G, CpE 493M, Mobile Robotics 6. Basic Probability Instructor: Yu Gu, Fall 2013 Uncertainties in Robotics Robot environments are inherently unpredictable; Sensors and data acquisition systems are
More informationMarkov Models and Reinforcement Learning. Stephen G. Ware CSCI 4525 / 5525
Markov Models and Reinforcement Learning Stephen G. Ware CSCI 4525 / 5525 Camera Vacuum World (CVW) 2 discrete rooms with cameras that detect dirt. A mobile robot with a vacuum. The goal is to ensure both
More informationORF 245 Fundamentals of Statistics Chapter 5 Probability
ORF 245 Fundamentals of Statistics Chapter 5 Probability Robert Vanderbei Oct 2015 Slides last edited on October 14, 2015 http://www.princeton.edu/ rvdb Sample Spaces (aka Populations) and Events When
More informationvalue of the sum standard units
Stat 1001 Winter 1998 Geyer Homework 7 Problem 18.1 20 and 25. Problem 18.2 (a) Average of the box. (1+3+5+7)=4=4. SD of the box. The deviations from the average are,3,,1, 1, 3. The squared deviations
More informationQuantitative Evaluation of Emedded Systems Solution 2: Discrete Time Markov Chains (DTMC)
Quantitative Evaluation of Emedded Systems Solution 2: Discrete Time Markov Chains (DTMC) 2.1 Classification of DTMC states Prof. Dr. Anne Remke Design and Analysis of Communication Systems University
More informationL09. PARTICLE FILTERING. NA568 Mobile Robotics: Methods & Algorithms
L09. PARTICLE FILTERING NA568 Mobile Robotics: Methods & Algorithms Particle Filters Different approach to state estimation Instead of parametric description of state (and uncertainty), use a set of state
More informationDiscrete Mathematics and Probability Theory Spring 2016 Rao and Walrand Note 14
CS 70 Discrete Mathematics and Probability Theory Spring 2016 Rao and Walrand Note 14 Introduction One of the key properties of coin flips is independence: if you flip a fair coin ten times and get ten
More informationAn Informal Introduction to Statistics in 2h. Tim Kraska
An Informal Introduction to Statistics in 2h Tim Kraska Goal of this Lecture This is not a replacement for a proper introduction to probability and statistics Instead, it only tries to convey the very
More informationAdvanced time-series analysis (University of Lund, Economic History Department)
Advanced time-series analysis (University of Lund, Economic History Department) 30 Jan-3 February and 26-30 March 2012 Lecture 3 Monte Carlo simulations and Bootstrapping. 3.a. What is a Monte Carlo simulation?
More informationLongest runs in coin tossing. Teaching recursive formulae, asymptotic theorems and computer simulations
tmcs-karacsony 2011/11/20 19:51 page 261 #1 9/2 (2011), 261 274 Longest runs in coin tossing. Teaching recursive formulae, asymptotic theorems and computer simulations Zsolt Karácsony and Józsefné Libor
More informationFinal. Introduction to Artificial Intelligence. CS 188 Spring You have approximately 2 hours and 50 minutes.
CS 188 Spring 2014 Introduction to Artificial Intelligence Final You have approximately 2 hours and 50 minutes. The exam is closed book, closed notes except your two-page crib sheet. Mark your answers
More informationDECISIONS UNDER UNCERTAINTY
August 18, 2003 Aanund Hylland: # DECISIONS UNDER UNCERTAINTY Standard theory and alternatives 1. Introduction Individual decision making under uncertainty can be characterized as follows: The decision
More informationPartially Observable Markov Decision Processes (POMDPs) Pieter Abbeel UC Berkeley EECS
Partially Observable Markov Decision Processes (POMDPs) Pieter Abbeel UC Berkeley EECS Many slides adapted from Jur van den Berg Outline POMDPs Separation Principle / Certainty Equivalence Locally Optimal
More informationReasoning Under Uncertainty: Introduction to Probability
Reasoning Under Uncertainty: Introduction to Probability CPSC 322 Lecture 23 March 12, 2007 Textbook 9 Reasoning Under Uncertainty: Introduction to Probability CPSC 322 Lecture 23, Slide 1 Lecture Overview
More informationThe Complexity of Change
The Complexity of Change JAN VAN DEN HEUVEL UQ, Brisbane, 26 July 2016 Department of Mathematics London School of Economics and Political Science A classical puzzle: the 15-Puzzle 13 2 3 12 1 2 3 4 9 11
More informationThe Reinforcement Learning Problem
The Reinforcement Learning Problem Slides based on the book Reinforcement Learning by Sutton and Barto Formalizing Reinforcement Learning Formally, the agent and environment interact at each of a sequence
More informationCMPSCI 240: Reasoning about Uncertainty
CMPSCI 240: Reasoning about Uncertainty Lecture 2: Sets and Events Andrew McGregor University of Massachusetts Last Compiled: January 27, 2017 Outline 1 Recap 2 Experiments and Events 3 Probabilistic Models
More information