The Bayes Theorema. Converting pre-diagnostic odds into post-diagnostic odds. Prof. Dr. F. Vanstapel, MD PhD Laboratoriumgeneeskunde UZ KULeuven
|
|
- Linda Freeman
- 5 years ago
- Views:
Transcription
1 slide 1 The Bayes Theorema Converting pre-diagnostic odds into post-diagnostic odds Prof. Dr. F. Vanstapel, MD PhD Laboratoriumgeneeskunde UZ KULeuven
2 slide 2 Problem * The yearly incidence of TBC infections in the school population of inner cities is 0.5 %. We tested your child radiographically and found a suspicious lesion. The test is believed to detect 80 % of early lesions. However 5% of healthy children show lesions which turn out to be benign artifacts. Are you alarmed? Do you have all the information that you need to have an idea of the chance that the suspicion turns out to be a serious condition? O Try to calculate the odds, or explain why you cannot * The example has no bearing with the actual situation in a particular poulation
3 slide 3 Solving the problem: 1/4 Given: Yearly incidence of TBC infections in the school population of inner cities is 0.5 %. P(D+) = We tested your child radiographically and found a suspicious lesion. The test is believed to detect 80 % of early lesions. P(T+ D+) = 0.80 However 5% of healthy subjects show lesions which turn out to be benign artifacts. P(T+ D-) = 0.05 What has to be solved: The probability that your child is sick, given the positive test result. P(D+ T+) Notation: P(D+) : the probability to find the disease in the population under study P(T+ D+) : the probability that the test is positive, when disease is present P(T+ D-) : the probability that the test is positive, when disease is absent P(D+ T+) : the probability to find the disease in persons with a positive test result
4 slide 4 Solving the problem: 2/4 Given: P(D+) = P(T+ D+) = 0.80 P(T+ D-) = 0.05 Solution: From known probabilities (proportions) TruePos = P(D+) * P(T+ D+) = * 0.80 = FalsePos = P(D-) * P(T+ D-) = ( )* P(D+ T+) = TruePos / (TruePos + FalsePos) / ( ) = 7.4 % P(D+ T+) P(D+) P(T+ D+) = 1- P(D+ T+) P(D-) P(T+ D-) Prevalence Sensitivity 1 - Specificity T+ T- T+ SUM True positive False positive D D D D Post-Test Odds P(D+ T+) 4/ /53.75
5 slide 5 Solving the problem: 3/4 Given: P(D+) = P(T+ D+) = 0.80 P(T+ D-) = 0.05 T+ T- D+ D Solution: From known odds (likelihood ratio s) PreTOdds = P(D+) / P(D-) = / ( ) = 1/199 LHR = P(T+ D+) / P(T+ D-) = 0.80 / 0.05 = 16/1 PostTOdds = PreTOdds * LHR = 16/199 P(D+ T+) = PreTOdds / (1 + PostTOdds) = 16 / ( ) = 7.4 % P(D+ T+) P(D+) P(T+ D+) = 1- P(D+ T+) P(D-) P(T+ D-) T+ D Pre-Test Odds Positive Likelihood Ratio Post-Test Odds P(D+ T+) D- Odds /199 16/1 16/199 16/215
6 slide 6 Test Result S S Latin square Pre-Test Probability Condition Prevalence: D/Total D _ TN FP TN / D _ D FN TP TP / D Specificity Sensitivity Test Ability P(S D) P(D) P(S D) TN / S TP / S Post-Test Probability Predictive Value NPV PPV P(D S) P(D S) What the clinician is after What the literature has on offer
7 slide 7 Solving the problem: 4/4 Given: P(D+) = Prevalence P(T+ D+) = Sensitivity P(T+ D-) = 1- Specificity Discussion of the solution: Bayes theorema solves P(D+ T+) from P(T+ D+) & P(T+ D-) & P(D+) P(D+ T+) P(D+) P(T+ D+) = = OddsRatio 1 - P(D+ T+) P(D-) P(T+ D-) OddsRatio P(D+ T+) = = Probability 1 + OddsRatio Prevalence, Sensitivity & Specificity is needed and suffices to solve the problem. The solution can be derived from probabilities (sol. 1) or from odds (sol. 2), with exactly the same number of numerical operations (starting from a Latin square).
8 Bayes Theorema - Introduction Version slide 8 Solving the problem - Comments (1/5) Equivalence of Proportions and Odds ratio s I got 1/8 of the cake Proportions Odds ratio s p all Formula s O = p / (1 - p) p = O / (1 + O) possibilities Break-Even-Point 1/2 1/1 Range 0 1/ /7 O all alternative possibilities %-transformed 0% 12.5% 100% log-transformed
9 Bayes Theorema - Introduction Version slide 9 Solving the problem - Comments (2/5) Likelihood ratio s : what they mean An example from the physical world : Dual-phase Extraction Odds Ratio s Proportions Phase 2 Vol 1 Vol 2 x Conc 1 Conc 2 = Mass 1 Mass 2 Mass 1 Mass 1 + Mass 2 yield partition coefficient Phase 1 phase volume ratio Likelihood ratio of an individual molecule at the interface to reside in either phase
10 Bayes Theorema - Introduction Version slide 10 Solving the problem - Comments (3/5) Dichotomic Classification versus a cut-off Unit Surfaces Chance Density x prevalence Frequency Distribution Number Density D- D+ D- Note: Optimal Cut-Off is Prevalence dependent D Bayesian Evaluation of Results Sensitivity Specificity P (T + D + ) P (T - D - ) Pre-test prevalence P (D + ) P (T + D + ) P(D + ) P (D - ) P (T - D - ) P(D - ) P (D + T + ) +
11 Bayes Theorema - Introduction Version slide 11 Solving the problem - Comments (4/5) Interpreting an actual value in a continuum of possible values Unit Surfaces Chance Density x prevalence Frequency Distribution Number Density D- D+ D- D Bayesian Evaluation of Results x Pre-test Prob P (x D + ) P (D + ) P (x D + ) P(D + ) P (D + x) + P (x D - ) P (D - ) P (x D - ) P(D - )
12 Bayes Theorema - Introduction Version slide 12 Solving the problem - Comments (5/5) Calculating with Odds Ratio s Pre-test Odds P (x D + ) P (x D - ) x P (D + ) = P (x D + ) P(D + ) P (D - ) P (x D - ) P(D - ) P (D + x) Post-test Odds non-linear parameter Prevalence in actual clinical case-mix in similar clinical presentations Positive Likelihood Ratio : non-linear parameter referring to artificial 1/1 prevalence
13 Bayes Theorema - Introduction Version slide 13 Summary: The Bayes theorema describes the learning process. It explains how prior knowledge is converted by a learning experience into posterior probability (= a considered professional judgement about the likelihood of a condition in a patient and about the diagnostic ability of the test requested) (= learning that a particular test result was obtained) (= degree of belief in a given diagnosis). The value of a test depends on its position in the sequence of diagnostic procedures The posterior probability (when expressed as such) is read on a scale of 0 to 1. The resulting decision is read on an on/off scale, and is either 0 or 1. Test drive our diagnostic calculator
14 slide 14 Concepts that you should have mastered in this exercise: O Prevalence O Pre-test Probability O Sensitivity and Specificity O Positive and (by analogy) Negative Likelihood Ratio s O Post-test Probability O Equivalence of Probabilities and Odds Exercise: Give for each item in the above list an operational (what you use it for) definition, and cross-check your definition with the solution of the problem. Exercise: In this example the test was used for screening purposes, sensitivity and specificity are defined with respect to a healthy population. What should be the definition of sensitivity and specificity in a clinical setting? (hint: differential diagnosis)
Bayesian Updating: Discrete Priors: Spring
Bayesian Updating: Discrete Priors: 18.05 Spring 2017 http://xkcd.com/1236/ Learning from experience Which treatment would you choose? 1. Treatment 1: cured 100% of patients in a trial. 2. Treatment 2:
More informationProbability and Statistics. Terms and concepts
Probability and Statistics Joyeeta Dutta Moscato June 30, 2014 Terms and concepts Sample vs population Central tendency: Mean, median, mode Variance, standard deviation Normal distribution Cumulative distribution
More informationBayesian Updating: Discrete Priors: Spring
Bayesian Updating: Discrete Priors: 18.05 Spring 2017 http://xkcd.com/1236/ Learning from experience Which treatment would you choose? 1. Treatment 1: cured 100% of patients in a trial. 2. Treatment 2:
More informationProbability and Statistics. Joyeeta Dutta-Moscato June 29, 2015
Probability and Statistics Joyeeta Dutta-Moscato June 29, 2015 Terms and concepts Sample vs population Central tendency: Mean, median, mode Variance, standard deviation Normal distribution Cumulative distribution
More informationLecture 18: Noise modeling and introduction to decision theory
Lecture 8: oise modeling and introduction to decision theory Learning Objectives: Hypothesis testing The receiver operator characteristic (ROC curve Bayes s Theorem, positive and negative predictive value
More informationReview. More Review. Things to know about Probability: Let Ω be the sample space for a probability measure P.
1 2 Review Data for assessing the sensitivity and specificity of a test are usually of the form disease category test result diseased (+) nondiseased ( ) + A B C D Sensitivity: is the proportion of diseased
More informationBase Rates and Bayes Theorem
Slides to accompany Grove s handout March 8, 2016 Table of contents 1 Diagnostic and Prognostic Inference 2 Example inferences Does this patient have Alzheimer s disease, schizophrenia, depression, etc.?
More informationConditional Probability. CS231 Dianna Xu
Conditional Probability CS231 Dianna Xu 1 Boy or Girl? A couple has two children, one of them is a girl. What is the probability that the other one is also a girl? Assuming 50/50 chances of conceiving
More informationBayesian Decision Theory
Introduction to Pattern Recognition [ Part 4 ] Mahdi Vasighi Remarks It is quite common to assume that the data in each class are adequately described by a Gaussian distribution. Bayesian classifier is
More informationA.I. in health informatics lecture 2 clinical reasoning & probabilistic inference, I. kevin small & byron wallace
A.I. in health informatics lecture 2 clinical reasoning & probabilistic inference, I kevin small & byron wallace today a review of probability random variables, maximum likelihood, etc. crucial for clinical
More informationDiagnostics. Gad Kimmel
Diagnostics Gad Kimmel Outline Introduction. Bootstrap method. Cross validation. ROC plot. Introduction Motivation Estimating properties of an estimator. Given data samples say the average. x 1, x 2,...,
More informationFormal Modeling in Cognitive Science
Formal Modeling in Cognitive Science Lecture 9: Application of Bayes Theorem; Discrete Random Variables; Steve Renals (notes by Frank Keller) School of Informatics University of Edinburgh s.renals@ed.ac.uk
More informationFormal Modeling in Cognitive Science Lecture 19: Application of Bayes Theorem; Discrete Random Variables; Distributions. Background.
Formal Modeling in Cognitive Science Lecture 9: ; Discrete Random Variables; Steve Renals (notes by Frank Keller) School of Informatics University of Edinburgh s.renals@ed.ac.uk February 7 Probability
More informationBayesian Updating: Discrete Priors: Spring 2014
ian Updating: Discrete Priors: 18.05 Spring 2014 http://xkcd.com/1236/ January 1, 2017 1 / 16 Learning from experience Which treatment would you choose? 1. Treatment 1: cured 100% of patients in a trial.
More informationMeasurement Systems. Biomedical Instrumentation. Biomedical Instrumentation. Lecture 3. Biomedical Instrumentation. Biomedical Instrumentation System
Biomedical Instrumentation Biomedical Instrumentation Prof. Dr. Nizamettin AYDIN naydin@yildiz.edu.tr naydin@ieee.org http://www.yildiz.edu.tr/~naydin Lecture 3 Measurement Systems 1 2 Biomedical Instrumentation
More informationApplied Machine Learning Annalisa Marsico
Applied Machine Learning Annalisa Marsico OWL RNA Bionformatics group Max Planck Institute for Molecular Genetics Free University of Berlin 22 April, SoSe 2015 Goals Feature Selection rather than Feature
More informationBayes Formula. MATH 107: Finite Mathematics University of Louisville. March 26, 2014
Bayes Formula MATH 07: Finite Mathematics University of Louisville March 26, 204 Test Accuracy Conditional reversal 2 / 5 A motivating question A rare disease occurs in out of every 0,000 people. A test
More informationFundamentals to Biostatistics. Prof. Chandan Chakraborty Associate Professor School of Medical Science & Technology IIT Kharagpur
Fundamentals to Biostatistics Prof. Chandan Chakraborty Associate Professor School of Medical Science & Technology IIT Kharagpur Statistics collection, analysis, interpretation of data development of new
More informationProbability. Introduction to Biostatistics
Introduction to Biostatistics Probability Second Semester 2014/2015 Text Book: Basic Concepts and Methodology for the Health Sciences By Wayne W. Daniel, 10 th edition Dr. Sireen Alkhaldi, BDS, MPH, DrPH
More information6.873/HST.951 Medical Decision Support Spring 2004 Evaluation
Harvard-MIT Division of Health Sciences and Technology HST.951J: Medical Decision Support, Fall 2005 Instructors: Professor Lucila Ohno-Machado and Professor Staal Vinterbo 6.873/HST.951 Medical Decision
More informationBayesian Networks. Probability Theory and Probabilistic Reasoning. Emma Rollon and Javier Larrosa Q
Bayesian Networks Probability Theory and Probabilistic Reasoning Emma Rollon and Javier Larrosa Q1-2015-2016 Emma Rollon and Javier Larrosa Bayesian Networks Q1-2015-2016 1 / 26 Degrees of belief Given
More informationProbability and Probability Distributions. Dr. Mohammed Alahmed
Probability and Probability Distributions 1 Probability and Probability Distributions Usually we want to do more with data than just describing them! We might want to test certain specific inferences about
More informationConditional Probabilities
Lecture Outline BIOST 514/517 Biostatistics I / pplied Biostatistics I Kathleen Kerr, Ph.D. ssociate Professor of Biostatistics University of Washington Probability Diagnostic Testing Random variables:
More informationFundamentals of Machine Learning for Predictive Data Analytics
Fundamentals of Machine Learning for Predictive Data Analytics Chapter 6: Probability-based Learning Sections 6.1, 6.2, 6.3 John Kelleher and Brian Mac Namee and Aoife D Arcy john.d.kelleher@dit.ie brian.macnamee@ucd.ie
More informationMachine Learning Linear Classification. Prof. Matteo Matteucci
Machine Learning Linear Classification Prof. Matteo Matteucci Recall from the first lecture 2 X R p Regression Y R Continuous Output X R p Y {Ω 0, Ω 1,, Ω K } Classification Discrete Output X R p Y (X)
More informationLecture 01: Introduction
Lecture 01: Introduction Dipankar Bandyopadhyay, Ph.D. BMTRY 711: Analysis of Categorical Data Spring 2011 Division of Biostatistics and Epidemiology Medical University of South Carolina Lecture 01: Introduction
More informationA.I. in health informatics lecture 3 clinical reasoning & probabilistic inference, II *
A.I. in health informatics lecture 3 clinical reasoning & probabilistic inference, II * kevin small & byron wallace * Slides borrow heavily from Andrew Moore, Weng- Keen Wong and Longin Jan Latecki today
More informationSolution to HW 12. Since B and B 2 form a partition, we have P (A) = P (A B 1 )P (B 1 ) + P (A B 2 )P (B 2 ). Using P (A) = 21.
Solution to HW 12 (1) (10 pts) Sec 12.3 Problem A screening test for a disease shows a positive result in 92% of all cases when the disease is actually present and in 7% of all cases when it is not. Assume
More informationExperimental Design and Data Analysis for Biologists
Experimental Design and Data Analysis for Biologists Gerry P. Quinn Monash University Michael J. Keough University of Melbourne CAMBRIDGE UNIVERSITY PRESS Contents Preface page xv I I Introduction 1 1.1
More informationLeast Squares Classification
Least Squares Classification Stephen Boyd EE103 Stanford University November 4, 2017 Outline Classification Least squares classification Multi-class classifiers Classification 2 Classification data fitting
More informationBayesian Learning Extension
Bayesian Learning Extension This document will go over one of the most useful forms of statistical inference known as Baye s Rule several of the concepts that extend from it. Named after Thomas Bayes this
More informationChapter 9: Hypothesis Testing Sections
1 / 22 : Hypothesis Testing Sections Skip: 9.2 Testing Simple Hypotheses Skip: 9.3 Uniformly Most Powerful Tests Skip: 9.4 Two-Sided Alternatives 9.5 The t Test 9.6 Comparing the Means of Two Normal Distributions
More informationAn AI-ish view of Probability, Conditional Probability & Bayes Theorem
An AI-ish view of Probability, Conditional Probability & Bayes Theorem Review: Uncertainty and Truth Values: a mismatch Let action A t = leave for airport t minutes before flight. Will A 15 get me there
More information10/18/2017. An AI-ish view of Probability, Conditional Probability & Bayes Theorem. Making decisions under uncertainty.
An AI-ish view of Probability, Conditional Probability & Bayes Theorem Review: Uncertainty and Truth Values: a mismatch Let action A t = leave for airport t minutes before flight. Will A 15 get me there
More informationQ1 (12 points): Chap 4 Exercise 3 (a) to (f) (2 points each)
Q1 (1 points): Chap 4 Exercise 3 (a) to (f) ( points each) Given a table Table 1 Dataset for Exercise 3 Instance a 1 a a 3 Target Class 1 T T 1.0 + T T 6.0 + 3 T F 5.0-4 F F 4.0 + 5 F T 7.0-6 F T 3.0-7
More informationModel Averaging (Bayesian Learning)
Model Averaging (Bayesian Learning) We want to predict the output Y of a new case that has input X = x given the training examples e: p(y x e) = m M P(Y m x e) = m M P(Y m x e)p(m x e) = m M P(Y m x)p(m
More informationFrequentist Statistics and Hypothesis Testing Spring
Frequentist Statistics and Hypothesis Testing 18.05 Spring 2018 http://xkcd.com/539/ Agenda Introduction to the frequentist way of life. What is a statistic? NHST ingredients; rejection regions Simple
More informationIntelligent Systems Statistical Machine Learning
Intelligent Systems Statistical Machine Learning Carsten Rother, Dmitrij Schlesinger WS2015/2016, Our model and tasks The model: two variables are usually present: - the first one is typically discrete
More informationBayes Classifiers. CAP5610 Machine Learning Instructor: Guo-Jun QI
Bayes Classifiers CAP5610 Machine Learning Instructor: Guo-Jun QI Recap: Joint distributions Joint distribution over Input vector X = (X 1, X 2 ) X 1 =B or B (drinking beer or not) X 2 = H or H (headache
More informationBayesian Concept Learning
Learning from positive and negative examples Bayesian Concept Learning Chen Yu Indiana University With both positive and negative examples, it is easy to define a boundary to separate these two. Just with
More informationUnit 2 Introduction to Probability
PubHlth 540 Fall 2013 2. Introduction to Probability Page 1 of 55 Unit 2 Introduction to Probability Chance favours only those who know how to court her - Charles Nicolle A weather report statement such
More informationLecture 10: Introduction to reasoning under uncertainty. Uncertainty
Lecture 10: Introduction to reasoning under uncertainty Introduction to reasoning under uncertainty Review of probability Axioms and inference Conditional probability Probability distributions COMP-424,
More informationCS 188: Artificial Intelligence. Our Status in CS188
CS 188: Artificial Intelligence Probability Pieter Abbeel UC Berkeley Many slides adapted from Dan Klein. 1 Our Status in CS188 We re done with Part I Search and Planning! Part II: Probabilistic Reasoning
More informationBasic Probabilistic Reasoning SEG
Basic Probabilistic Reasoning SEG 7450 1 Introduction Reasoning under uncertainty using probability theory Dealing with uncertainty is one of the main advantages of an expert system over a simple decision
More informationClassifier performance evaluation
Classifier performance evaluation Václav Hlaváč Czech Technical University in Prague Czech Institute of Informatics, Robotics and Cybernetics 166 36 Prague 6, Jugoslávských partyzánu 1580/3, Czech Republic
More informationNaive Bayes classification
Naive Bayes classification Christos Dimitrakakis December 4, 2015 1 Introduction One of the most important methods in machine learning and statistics is that of Bayesian inference. This is the most fundamental
More informationIntroduction to Supervised Learning. Performance Evaluation
Introduction to Supervised Learning Performance Evaluation Marcelo S. Lauretto Escola de Artes, Ciências e Humanidades, Universidade de São Paulo marcelolauretto@usp.br Lima - Peru Performance Evaluation
More informationPHP 2510 Probability Part 3. Calculate probability by counting (cont ) Conditional probability. Law of total probability. Bayes Theorem.
PHP 2510 Probability Part 3 Calculate probability by counting (cont ) Conditional probability Law of total probability Bayes Theorem Independence PHP 2510 Sept 22, 2009 1 Last time... Count the number
More informationStatistical Methods in Particle Physics. Lecture 2
Statistical Methods in Particle Physics Lecture 2 October 17, 2011 Silvia Masciocchi, GSI Darmstadt s.masciocchi@gsi.de Winter Semester 2011 / 12 Outline Probability Definition and interpretation Kolmogorov's
More informationIt applies to discrete and continuous random variables, and a mix of the two.
3. Bayes Theorem 3.1 Bayes Theorem A straightforward application of conditioning: using p(x, y) = p(x y) p(y) = p(y x) p(x), we obtain Bayes theorem (also called Bayes rule) p(x y) = p(y x) p(x). p(y)
More informationAnnouncements. Topics: To Do:
Announcements Topics: In the Probability and Statistics module: - Sections 1 + 2: Introduction to Stochastic Models - Section 3: Basics of Probability Theory - Section 4: Conditional Probability; Law of
More informationBayesian Networks Basic and simple graphs
Bayesian Networks Basic and simple graphs Ullrika Sahlin, Centre of Environmental and Climate Research Lund University, Sweden Ullrika.Sahlin@cec.lu.se http://www.cec.lu.se/ullrika-sahlin Bayesian [Belief]
More informationClassifier Evaluation. Learning Curve cleval testc. The Apparent Classification Error. Error Estimation by Test Set. Classifier
Classifier Learning Curve How to estimate classifier performance. Learning curves Feature curves Rejects and ROC curves True classification error ε Bayes error ε* Sub-optimal classifier Bayes consistent
More informationBANA 7046 Data Mining I Lecture 4. Logistic Regression and Classications 1
BANA 7046 Data Mining I Lecture 4. Logistic Regression and Classications 1 Shaobo Li University of Cincinnati 1 Partially based on Hastie, et al. (2009) ESL, and James, et al. (2013) ISLR Data Mining I
More informationMeta-Analysis for Diagnostic Test Data: a Bayesian Approach
Meta-Analysis for Diagnostic Test Data: a Bayesian Approach Pablo E. Verde Coordination Centre for Clinical Trials Heinrich Heine Universität Düsseldorf Preliminaries: motivations for systematic reviews
More informationCDA6530: Performance Models of Computers and Networks. Chapter 1: Review of Practical Probability
CDA6530: Performance Models of Computers and Networks Chapter 1: Review of Practical Probability Probability Definition Sample Space (S) which is a collection of objects (all possible scenarios or values).
More informationSpatially-Explicit Bayesian Information Entropy Metrics for Calibrating Landscape Transformation Models
Entropy 2013, 15, 2480-2509; doi:10.3390/e15072480 Article OPEN ACCESS entropy ISSN 1099-4300 www.mdpi.com/journal/entropy Spatially-Explicit Bayesian Information Entropy Metrics for Calibrating Landscape
More informationBayesian decision theory. Nuno Vasconcelos ECE Department, UCSD
Bayesian decision theory Nuno Vasconcelos ECE Department, UCSD Notation the notation in DHS is quite sloppy e.g. show that ( error = ( error z ( z dz really not clear what this means we will use the following
More informationContents. Decision Making under Uncertainty 1. Meanings of uncertainty. Classical interpretation
Contents Decision Making under Uncertainty 1 elearning resources Prof. Ahti Salo Helsinki University of Technology http://www.dm.hut.fi Meanings of uncertainty Interpretations of probability Biases in
More informationIntelligent Systems Statistical Machine Learning
Intelligent Systems Statistical Machine Learning Carsten Rother, Dmitrij Schlesinger WS2014/2015, Our tasks (recap) The model: two variables are usually present: - the first one is typically discrete k
More informationBiochip informatics-(i)
Biochip informatics-(i) : biochip normalization & differential expression Ju Han Kim, M.D., Ph.D. SNUBI: SNUBiomedical Informatics http://www.snubi snubi.org/ Biochip Informatics - (I) Biochip basics Preprocessing
More informationRecall from last time: Conditional probabilities. Lecture 2: Belief (Bayesian) networks. Bayes ball. Example (continued) Example: Inference problem
Recall from last time: Conditional probabilities Our probabilistic models will compute and manipulate conditional probabilities. Given two random variables X, Y, we denote by Lecture 2: Belief (Bayesian)
More informationStephen Scott.
1 / 35 (Adapted from Ethem Alpaydin and Tom Mitchell) sscott@cse.unl.edu In Homework 1, you are (supposedly) 1 Choosing a data set 2 Extracting a test set of size > 30 3 Building a tree on the training
More informationProbability. Lecture Notes. Adolfo J. Rumbos
Probability Lecture Notes Adolfo J. Rumbos October 20, 204 2 Contents Introduction 5. An example from statistical inference................ 5 2 Probability Spaces 9 2. Sample Spaces and σ fields.....................
More informationUsing Probability to do Statistics.
Al Nosedal. University of Toronto. November 5, 2015 Milk and honey and hemoglobin Animal experiments suggested that honey in a diet might raise hemoglobin level. A researcher designed a study involving
More informationError Rates. Error vs Threshold. ROC Curve. Biometrics: A Pattern Recognition System. Pattern classification. Biometrics CSE 190 Lecture 3
Biometrics: A Pattern Recognition System Yes/No Pattern classification Biometrics CSE 190 Lecture 3 Authentication False accept rate (FAR): Proportion of imposters accepted False reject rate (FRR): Proportion
More informationBayes Theorem & Diagnostic Tests Screening Tests
Bayes heore & Diagnostic ests Screening ests Box contains 2 red balls and blue ball Box 2 contains red ball and 3 blue balls A coin is tossed. If Head turns up a ball is drawn fro Box, and if ail turns
More informationStatistical methods for decision making in mine action
Statistical methods for decision making in mine action Jan Larsen Intelligent Signal Processing Technical University of Denmark jl@imm.dtu.dk, www.imm.dtu.dk/~jl Jan Larsen 1 Why do we need statistical
More informationUncertainty. Variables. assigns to each sentence numerical degree of belief between 0 and 1. uncertainty
Bayes Classification n Uncertainty & robability n Baye's rule n Choosing Hypotheses- Maximum a posteriori n Maximum Likelihood - Baye's concept learning n Maximum Likelihood of real valued function n Bayes
More informationIntermediate Math Circles November 15, 2017 Probability III
Intermediate Math Circles November 5, 07 Probability III Example : You have bins in which there are coloured balls. The balls are identical except for their colours. The contents of the containers are:
More informationLecture 3. STAT161/261 Introduction to Pattern Recognition and Machine Learning Spring 2018 Prof. Allie Fletcher
Lecture 3 STAT161/261 Introduction to Pattern Recognition and Machine Learning Spring 2018 Prof. Allie Fletcher Previous lectures What is machine learning? Objectives of machine learning Supervised and
More informationL2: Review of probability and statistics
Probability L2: Review of probability and statistics Definition of probability Axioms and properties Conditional probability Bayes theorem Random variables Definition of a random variable Cumulative distribution
More informationIntroduction to Artificial Intelligence. Prof. Inkyu Moon Dept. of Robotics Engineering, DGIST
Introduction to Artificial Intelligence Prof. Inkyu Moon Dept. of Robotics Engineering, DGIST Chapter 3 Uncertainty management in rule-based expert systems To help interpret Bayesian reasoning in expert
More information6 Event-based Independence and Conditional Probability. Sneak peek: Figure 3: Conditional Probability Example: Sneak Peek
6 Event-based Independence and Conditional Probability Example Example Roll 6.1. a fair Roll dicea dice... Sneak peek: Figure 3: Conditional Probability Example: Sneak Peek Example 6.2 (Slides). Diagnostic
More informationBayesian Updating with Discrete Priors Class 11, Jeremy Orloff and Jonathan Bloom
1 Learning Goals ian Updating with Discrete Priors Class 11, 18.05 Jeremy Orloff and Jonathan Bloom 1. Be able to apply theorem to compute probabilities. 2. Be able to define the and to identify the roles
More informationPerformance evaluation of binary classifiers
Performance evaluation of binary classifiers Kevin P. Murphy Last updated October 10, 2007 1 ROC curves We frequently design systems to detect events of interest, such as diseases in patients, faces in
More informationMath for Machine Learning Open Doors to Data Science and Artificial Intelligence. Richard Han
Math for Machine Learning Open Doors to Data Science and Artificial Intelligence Richard Han Copyright 05 Richard Han All rights reserved. CONTENTS PREFACE... - INTRODUCTION... LINEAR REGRESSION... 4 LINEAR
More informationBayesian Inference. Introduction
Bayesian Inference Introduction The frequentist approach to inference holds that probabilities are intrinsicially tied (unsurprisingly) to frequencies. This interpretation is actually quite natural. What,
More informationCLASSIFICATION NAIVE BAYES. NIKOLA MILIKIĆ UROŠ KRČADINAC
CLASSIFICATION NAIVE BAYES NIKOLA MILIKIĆ nikola.milikic@fon.bg.ac.rs UROŠ KRČADINAC uros@krcadinac.com WHAT IS CLASSIFICATION? A supervised learning task of determining the class of an instance; it is
More informationPerformance Evaluation and Comparison
Outline Hong Chang Institute of Computing Technology, Chinese Academy of Sciences Machine Learning Methods (Fall 2012) Outline Outline I 1 Introduction 2 Cross Validation and Resampling 3 Interval Estimation
More informationMachine Learning for natural language processing
Machine Learning for natural language processing Classification: Naive Bayes Laura Kallmeyer Heinrich-Heine-Universität Düsseldorf Summer 2016 1 / 20 Introduction Classification = supervised method for
More informationFinal Examination CS540-2: Introduction to Artificial Intelligence
Final Examination CS540-2: Introduction to Artificial Intelligence May 9, 2018 LAST NAME: SOLUTIONS FIRST NAME: Directions 1. This exam contains 33 questions worth a total of 100 points 2. Fill in your
More informationJoint, Conditional, & Marginal Probabilities
Joint, Conditional, & Marginal Probabilities The three axioms for probability don t discuss how to create probabilities for combined events such as P [A B] or for the likelihood of an event A given that
More informationProbability, Statistics, and Bayes Theorem Session 3
Probability, Statistics, and Bayes Theorem Session 3 1 Introduction Now that we know what Bayes Theorem is, we want to explore some of the ways that it can be used in real-life situations. Often the results
More informationLecture 3 Classification, Logistic Regression
Lecture 3 Classification, Logistic Regression Fredrik Lindsten Division of Systems and Control Department of Information Technology Uppsala University. Email: fredrik.lindsten@it.uu.se F. Lindsten Summary
More informationProbabilistic Graphical Models for Image Analysis - Lecture 1
Probabilistic Graphical Models for Image Analysis - Lecture 1 Alexey Gronskiy, Stefan Bauer 21 September 2018 Max Planck ETH Center for Learning Systems Overview 1. Motivation - Why Graphical Models 2.
More informationPattern Classification
Pattern Classification All materials in these slides were taken from Pattern Classification (2nd ed) by R. O. Duda, P. E. Hart and D. G. Stork, John Wiley & Sons, 2000 with the permission of the authors
More informationIntrusion Detection: Base Rate Fallacy
CS 5323 Intrusion Detection: Base ate Fallacy Prof. avi Sandhu Executive Director and Endowed Chair Lecture 11 ravi.utsa@gmail.com www.profsandhu.com avi Sandhu 1 Base-ate Fallacy ᴧ S ᴧ S : Test esult
More informationChapter 6. Probability
Chapter 6 robability Suppose two six-sided die is rolled and they both land on sixes. Or a coin is flipped and it lands on heads. Or record the color of the next 20 cars to pass an intersection. These
More informationWith Question/Answer Animations. Chapter 7
With Question/Answer Animations Chapter 7 Chapter Summary Introduction to Discrete Probability Probability Theory Bayes Theorem Section 7.1 Section Summary Finite Probability Probabilities of Complements
More informationBayesian data analysis using JASP
Bayesian data analysis using JASP Dani Navarro compcogscisydney.com/jasp-tute.html Part 1: Theory Philosophy of probability Introducing Bayes rule Bayesian reasoning A simple example Bayesian hypothesis
More informationCOMP 551 Applied Machine Learning Lecture 19: Bayesian Inference
COMP 551 Applied Machine Learning Lecture 19: Bayesian Inference Associate Instructor: (herke.vanhoof@mcgill.ca) Class web page: www.cs.mcgill.ca/~jpineau/comp551 Unless otherwise noted, all material posted
More information4. Conditional Probability
1 of 13 7/15/2009 9:25 PM Virtual Laboratories > 2. Probability Spaces > 1 2 3 4 5 6 7 4. Conditional Probability Definitions and Interpretations The Basic Definition As usual, we start with a random experiment
More informationBayesian Statistics Part III: Building Bayes Theorem Part IV: Prior Specification
Bayesian Statistics Part III: Building Bayes Theorem Part IV: Prior Specification Michael Anderson, PhD Hélène Carabin, DVM, PhD Department of Biostatistics and Epidemiology The University of Oklahoma
More informationLearning From Crowds. Presented by: Bei Peng 03/24/15
Learning From Crowds Presented by: Bei Peng 03/24/15 1 Supervised Learning Given labeled training data, learn to generalize well on unseen data Binary classification ( ) Multi-class classification ( y
More informationLogistics. Naïve Bayes & Expectation Maximization. 573 Schedule. Coming Soon. Estimation Models. Topics
Logistics Naïve Bayes & Expectation Maximization CSE 7 eam Meetings Midterm Open book, notes Studying See AIMA exercises Daniel S. Weld Daniel S. Weld 7 Schedule Selected opics Coming Soon Selected opics
More informationData Mining: Concepts and Techniques. (3 rd ed.) Chapter 8. Chapter 8. Classification: Basic Concepts
Data Mining: Concepts and Techniques (3 rd ed.) Chapter 8 Chapter 8. Classification: Basic Concepts Classification: Basic Concepts Decision Tree Induction Bayes Classification Methods Rule-Based Classification
More informationData Mining: Concepts and Techniques. (3 rd ed.) Chapter 8. Chapter 8. Classification: Basic Concepts
Data Mining: Concepts and Techniques (3 rd ed.) Chapter 8 1 Chapter 8. Classification: Basic Concepts Classification: Basic Concepts Decision Tree Induction Bayes Classification Methods Rule-Based Classification
More informationBayesian Logic Thomas Bayes:
Bayesian Logic Thomas Bayes: 1702-1761 Based on probability, P(X) NOT frequency distributions, BUT all common distributions can be incorporated into a Bayesian analysis : Poisson, Normal (Gaussian) etc.
More informationConfusion matrix. a = true positives b = false negatives c = false positives d = true negatives 1. F-measure combines Recall and Precision:
Confusion matrix classifier-determined positive label classifier-determined negative label true positive a b label true negative c d label Accuracy = (a+d)/(a+b+c+d) a = true positives b = false negatives
More information