Formal Modeling in Cognitive Science

Similar documents
Formal Modeling in Cognitive Science Lecture 19: Application of Bayes Theorem; Discrete Random Variables; Distributions. Background.

Mathematics. ( : Focus on free Education) (Chapter 16) (Probability) (Class XI) Exercise 16.2

Conditional Probability

Why should you care?? Intellectual curiosity. Gambling. Mathematically the same as the ESP decision problem we discussed in Week 4.

SDS 321: Introduction to Probability and Statistics

Bayesian Updating: Discrete Priors: Spring

Outline. 1. Define likelihood 2. Interpretations of likelihoods 3. Likelihood plots 4. Maximum likelihood 5. Likelihood ratio benchmarks

Bayesian Updating: Discrete Priors: Spring

MODULE 2 RANDOM VARIABLE AND ITS DISTRIBUTION LECTURES DISTRIBUTION FUNCTION AND ITS PROPERTIES

First Digit Tally Marks Final Count

Deep Learning for Computer Vision

Probability deals with modeling of random phenomena (phenomena or experiments whose outcomes may vary)

Expected Value. Lecture A Tiefenbruck MWF 9-9:50am Center 212 Lecture B Jones MWF 2-2:50pm Center 214 Lecture C Tiefenbruck MWF 11-11:50am Center 212

CSC Discrete Math I, Spring Discrete Probability

Lecture 6 Random Variable. Compose of procedure & observation. From observation, we get outcomes

CS4705. Probability Review and Naïve Bayes. Slides from Dragomir Radev

V. RANDOM VARIABLES, PROBABILITY DISTRIBUTIONS, EXPECTED VALUE

Probability and Probability Distributions. Dr. Mohammed Alahmed

Outline. 1. Define likelihood 2. Interpretations of likelihoods 3. Likelihood plots 4. Maximum likelihood 5. Likelihood ratio benchmarks

27 Binary Arithmetic: An Application to Programming

What is a random variable

Bayesian Updating: Discrete Priors: Spring 2014

Business Statistics MBA Pokhara University

Probability: Terminology and Examples Class 2, Jeremy Orloff and Jonathan Bloom

Decision Trees. Nicholas Ruozzi University of Texas at Dallas. Based on the slides of Vibhav Gogate and David Sontag

Lecture 3 - Axioms of Probability

Mean, Median and Mode. Lecture 3 - Axioms of Probability. Where do they come from? Graphically. We start with a set of 21 numbers, Sta102 / BME102

Conditional Probability

Computational Cognitive Science

(i) Given that a student is female, what is the probability of having a GPA of at least 3.0?

2. Conditional Probability

4/17/2012. NE ( ) # of ways an event can happen NS ( ) # of events in the sample space

Bayes Rule for probability

Lecture Notes 1 Basic Probability. Elements of Probability. Conditional probability. Sequential Calculation of Probability

Probability Distributions. Conditional Probability

Computational Cognitive Science

Probability Notes (A) , Fall 2010

A.I. in health informatics lecture 2 clinical reasoning & probabilistic inference, I. kevin small & byron wallace

Events A and B are said to be independent if the occurrence of A does not affect the probability of B.

ELEG 3143 Probability & Stochastic Process Ch. 1 Experiments, Models, and Probabilities

Introduction Probability. Math 141. Introduction to Probability and Statistics. Albyn Jones


Random Variables. Statistics 110. Summer Copyright c 2006 by Mark E. Irwin

Some Probability and Statistics

Discrete Mathematics and Probability Theory Spring 2014 Anant Sahai Note 10

Statistical methods in recognition. Why is classification a problem?

NLP: Probability. 1 Basics. Dan Garrette December 27, E : event space (sample space)

Steve Smith Tuition: Maths Notes

Basic Probability Theory (I)

Statistics Statistical Process Control & Control Charting

Bayes Formula. MATH 107: Finite Mathematics University of Louisville. March 26, 2014

Probability Distributions for Discrete RV

Probability Theory. Introduction to Probability Theory. Principles of Counting Examples. Principles of Counting. Probability spaces.

Chapter 2 PROBABILITY SAMPLE SPACE

PERMUTATIONS, COMBINATIONS AND DISCRETE PROBABILITY

Quantitative Methods for Decision Making

Tutorial for Lecture Course on Modelling and System Identification (MSI) Albert-Ludwigs-Universität Freiburg Winter Term

Lecture 1: Probability Fundamentals

Probability (10A) Young Won Lim 6/12/17

Lecture 3: Random variables, distributions, and transformations

Probability Pearson Education, Inc. Slide

9/6/2016. Section 5.1 Probability. Equally Likely Model. The Division Rule: P(A)=#(A)/#(S) Some Popular Randomizers.

Today we ll discuss ways to learn how to think about events that are influenced by chance.

Preliminary Statistics Lecture 2: Probability Theory (Outline) prelimsoas.webs.com

ECE 450 Lecture 2. Recall: Pr(A B) = Pr(A) + Pr(B) Pr(A B) in general = Pr(A) + Pr(B) if A and B are m.e. Lecture Overview

Basic Statistics for SGPE Students Part II: Probability theory 1

Machine Learning

Discrete Structures for Computer Science

Solution to HW 12. Since B and B 2 form a partition, we have P (A) = P (A B 1 )P (B 1 ) + P (A B 2 )P (B 2 ). Using P (A) = 21.

Lesson B1 - Probability Distributions.notebook

What is Probability? Probability. Sample Spaces and Events. Simple Event

Probability and Statistics. Joyeeta Dutta-Moscato June 29, 2015

TA Qinru Shi: Based on poll result, the first Programming Boot Camp will be: this Sunday 5 Feb, 7-8pm Gates 114.

6.2 Introduction to Probability. The Deal. Possible outcomes: STAT1010 Intro to probability. Definitions. Terms: What are the chances of?

MAE 493G, CpE 493M, Mobile Robotics. 6. Basic Probability

Computational Cognitive Science

STAT 430/510 Probability

Notation: X = random variable; x = particular value; P(X = x) denotes probability that X equals the value x.

Probability and Statistics. Terms and concepts

Notes slides from before lecture. CSE 21, Winter 2017, Section A00. Lecture 16 Notes. Class URL:

Probability Pr(A) 0, for any event A. 2. Pr(S) = 1, for the sample space S. 3. If A and B are mutually exclusive, Pr(A or B) = Pr(A) + Pr(B).

2. In a clinical trial of certain new treatment, we may be interested in the proportion of patients cured.

Review. More Review. Things to know about Probability: Let Ω be the sample space for a probability measure P.

Notes slides from before lecture. CSE 21, Winter 2017, Section A00. Lecture 15 Notes. Class URL:

Probability theory. References:

Priors, Total Probability, Expectation, Multiple Trials

Chapter 3 Questions. Question 3.1. Based on the nature of values that each random variable can take, we can have the following classifications:

L2: Review of probability and statistics

Discrete Random Variable

MATH 10 INTRODUCTORY STATISTICS

Conditional Probability and Bayes Theorem (2.4) Independence (2.5)

Lecture 1 : The Mathematical Theory of Probability

With Question/Answer Animations. Chapter 7

Dept. of Linguistics, Indiana University Fall 2015

Example A. Define X = number of heads in ten tosses of a coin. What are the values that X may assume?

COS 424: Interacting with Data. Lecturer: Dave Blei Lecture #2 Scribe: Ellen Kim February 7, 2008

Lecture 10: Introduction to reasoning under uncertainty. Uncertainty

Probability. Lecture Notes. Adolfo J. Rumbos

Probability. VCE Maths Methods - Unit 2 - Probability

Some Probability and Statistics

Transcription:

Formal Modeling in Cognitive Science Lecture 9: Application of Bayes Theorem; Discrete Random Variables; Steve Renals (notes by Frank Keller) School of Informatics University of Edinburgh s.renals@ed.ac.uk 22 February 2007 Steve Renals (notes by Frank Keller) Formal Modeling in Cognitive Science

Application of Bayes Theorem Background Application to Diagnosis Base Rate Neglect 2 3 Probability Cumulative Steve Renals (notes by Frank Keller) Formal Modeling in Cognitive Science 2

Background Application of Bayes Theorem Background Application to Diagnosis Base Rate Neglect Let s look at an application of Bayes theorem to the analysis of cognitive processes. First we need to introduce some data. Research on human decision making investigates, e.g., how physicians make a medical diagnosis (Casscells et al. 97): Example If a test to detect a disease whose prevalence is /000 has a false-positive rate of 5%, what is the chance that a person found to have a positive result actually has the disease, assuming you know nothing about the person s symptoms or signs? Steve Renals (notes by Frank Keller) Formal Modeling in Cognitive Science 3

Background Application of Bayes Theorem Background Application to Diagnosis Base Rate Neglect Most frequent answer: 95% Reasoning: if false-positive rate is 5%, then test will be correct 95% of the time. Correct answer: 2% Reasoning: assume you test 000 people; the test will be positive in 50 cases (5%), but only one person actually has the disease. Hence the chance that a person with a positive result has the disease is /50 = 2%. Only 2% of subjects give the correct answer. Mathematics underlying the correct answer: Bayes Theorem. Steve Renals (notes by Frank Keller) Formal Modeling in Cognitive Science 4

Bayes Theorem Application of Bayes Theorem Background Application to Diagnosis Base Rate Neglect We need to think about Bayes theorem slightly differently to apply it to this problem (and the terms have special names now): Bayes Theorem (for hypothesis testing) Given a hypothesis h and data D which bears on the hypothesis: P(h D) = P(D h)p(h) P(D) P(h): independent probability of h: prior probability P(D): independent probability of D P(D h): conditional probability of D given h: likelihood P(h D): conditional probability of h given D: posterior probability Steve Renals (notes by Frank Keller) Formal Modeling in Cognitive Science 5

Application to Diagnosis Background Application to Diagnosis Base Rate Neglect In Casscells et al. s (97) examples, we have the following: h: person tested has the disease; h: person tested doesn t have the disease; D: person tests positive for the disease. The following probabilities are known: P(h) = /000 = 0.00 P( h) = P(h) = 0.999 P(D h) = 5% = 0.05 P(D h) = (assume perfect test) Compute the probability of the data (rule of total probability): P(D) = P(D h)p(h)+p(d h)p( h) = 0.00+0.05 0.999 = 0.05095 Compute the probability of correctly detecting the illness: P(h D) = P(h)P(D h) P(D) = 0.00 0.05095 = 0.0963 Steve Renals (notes by Frank Keller) Formal Modeling in Cognitive Science 6

Base Rate Neglect Application of Bayes Theorem Background Application to Diagnosis Base Rate Neglect Base rate: the probability of the hypothesis being true in the absence of any data (i.e., prior probability). Base rate neglect: people have a tendency to ignore base rate information (see Casscells et al. s (97) experimental results). base rate neglect has been demonstrated in a number of experimental situations; often presented as a fundamental bias in decision making; however, experiments show that subjects use base rates in certain situations; it has been argued that base rate neglect is only occurs in artificial or abstract mathematical situations. Steve Renals (notes by Frank Keller) Formal Modeling in Cognitive Science 7

Base Rates and Experience Background Application to Diagnosis Base Rate Neglect Potential problems with in Casscells et al. s (97) study: subjects were simply told the statistical facts; they had no first-hand experience with the facts (through exposure to many applications of the test); providing subjects with experience has been shown to reduce or eliminate base rate neglect. Medin and Edelson (9) tested the role of experience on decision making in medical diagnosis. Steve Renals (notes by Frank Keller) Formal Modeling in Cognitive Science

Base Rates and Experience Background Application to Diagnosis Base Rate Neglect Medin and Edelson (9) trained subjects on a diagnosis task in which diseases varied in frequency: subjects were presented with pairs of symptoms and had to select one of six diseases; feedback was provided so that they learned symptom/disease associations; base rates of the diseases were manipulated; once subjects had achieved perfect diagnosis accuracy, they entered the transfer phase; subjects now made diagnoses for combinations of symptoms they had not seen before; made use of base rate information. Steve Renals (notes by Frank Keller) Formal Modeling in Cognitive Science 9

Definition: Random Variable If S is a sample space with a probability measure and X is a real-valued function defined over the elements of S, then X is called a random variable. We will denote random variable by capital letters (e.g., X ), and their values by lower-case letters (e.g., x). Example Given an experiment in which we roll a pair of dice, let the random variable X be the total number of points rolled with the two dice. For example X = 7 picks out the set {(, 6), (2, 5), (3, 4), (4, 3), (5, 2), (6, )}. Steve Renals (notes by Frank Keller) Formal Modeling in Cognitive Science 0

This can be illustrated graphically: 6 5 4 3 2 die 2 7 9 0 2 6 7 9 0 5 6 7 9 0 4 5 6 7 9 3 4 5 6 7 2 3 4 5 6 7 die 2 3 4 5 6 For each outcome, this graph lists the value of X, and the dotted area corresponds to X = 7. Steve Renals (notes by Frank Keller) Formal Modeling in Cognitive Science

Example Assume a balanced coin is flipped three times. Let X be the random variable denoting the total number of heads obtained. Outcome Probability x HHH 3 HHT 2 HTH 2 THH 2 Outcome Probability x TTH THT HTT TTT 0 Hence, P(X = 0) =, P(X = ) = P(X = 2) = 3, P(X = 3) =. Steve Renals (notes by Frank Keller) Formal Modeling in Cognitive Science 2

Probability Probability Cumulative Definition: Probability Distribution If X is a discrete random variable, the function given by f (x) = P(X = x) for each x within the range of X is called the probability distribution of X. Theorem: Probability Distribution A function can serve as the probability distribution of a discrete random variable X if and only if its values, f (x), satisfy the conditions: f (x) 0 for each value within its domain; 2 x f (x) =, where x over all the values within its domain. Steve Renals (notes by Frank Keller) Formal Modeling in Cognitive Science 3

Probability Probability Cumulative Example For the probability function defined in the previous example: x f (x) = P(X = x) 0 3 2 3 3 This function can be written more concisely as: f (x) = 4 3 2x Steve Renals (notes by Frank Keller) Formal Modeling in Cognitive Science 4

Probability Probability Cumulative A probability distribution is often represented as a probability histogram. For the previous example: 0.9 0. 0.7 0.6 f(x) 0.5 0.4 0.3 0.2 0. 0 0 2 3 x Steve Renals (notes by Frank Keller) Formal Modeling in Cognitive Science 5

Cumulative Distribution Probability Cumulative In many cases, we re interested in the probability for values X x, rather than for X = x. Definition: Cumulative Distribution If X is a discrete random variable, the function given by: F (x) = P(X x) = t x f (t) for < x < where f (t) is the value of the probability distribution of X at t, is the cumulative distribution of X. Steve Renals (notes by Frank Keller) Formal Modeling in Cognitive Science 6

Cumulative Probability Cumulative Example Consider the probability distribution f (x) = 4 3 2x from the previous example. The values of the cumulative distribution are: F (x) = x f (x) F(x) 0 3 3 2 3 4 7 Note that F (x) is defined for all real values of x: 0 for x < 0 for 0 x < 4 for x < 2 7 for 2 x < 3 for x 3 Steve Renals (notes by Frank Keller) Formal Modeling in Cognitive Science 7

Cumulative Probability Cumulative The cumulative distribution can be graphed; for the previous example: F(x) 0.9 0. 0.7 0.6 0.5 0.4 0.3 0.2 0. 0 0 2 3 4 x Steve Renals (notes by Frank Keller) Formal Modeling in Cognitive Science

Cumulative Probability Cumulative Theorem: Cumulative The values F (x) of the cumulative distribution of a discrete random variable X satisfies the conditions: F ( ) = 0 and F ( ) = ; 2 if a < b, then F (a) F (b) for any real numbers a and b. Example Consider the example of F (x) on the previous slide: F ( ) = 0 as F (0) < 0 by definition; F ( ) = as F ( ) 3 by definition; 2 F (a) < F (b) holds for (0, ), (, 2), (2, 3) by definition; F (a) = F (b) holds for all other values of a and b. Steve Renals (notes by Frank Keller) Formal Modeling in Cognitive Science 9

Summary Application of Bayes Theorem Probability Cumulative There are many applications of Bayes theorem in cognitive science (here: medical diagnosis); base rate neglect: experimental subjects ignore information about prior probability; a random variable picks out a subset of the sample space; a probability distribution returns a probability for each value of a random variable. a cumulative distribution sums all the values of a probability up to a threshold. Steve Renals (notes by Frank Keller) Formal Modeling in Cognitive Science 20

References Application of Bayes Theorem Probability Cumulative Casscells, W., A. Schoenberger, and T. Grayboys. 97. Interpretation by physicians of clinical laboratory results. New England Journal of Medicine 299():999 00. Medin, D. L. and S. M. Edelson. 9. Problem structure and the use of base-rate information from experience. Journal of Experimental Psychology: General 7():6 5. Steve Renals (notes by Frank Keller) Formal Modeling in Cognitive Science 2