12/2/15. G Perception. Bayesian Decision Theory. Laurence T. Maloney. Perceptual Tasks. Testing hypotheses. Estimation
|
|
- Morris Murphy
- 5 years ago
- Views:
Transcription
1 G Perception Bayesian Decision Theory Laurence T. Maloney Perceptual Tasks Testing hypotheses signal detection theory psychometric function Estimation previous lecture Selection of actions this lecture Criteria Estimation Maximize information Selection of actions Maximize reward (value) 1
2 What is the action that maximizes your expected reward (value)? Statistical decision theory is a model of action selection Statistical Decision Theory Abraham Wald John von Neumann David Blackwell Oskar Morgenstern M. A. Girschick 1954 The Three Elements of SDT W A X { } { } { } = w 1,w 2,...,w m = a 1,a 2,...,a p = x 1,x 2...,x n possible states of the world possible actions possible sensory events 2
3 Bayesian Decision Theory (BDT) gain G( a, w ) Consequence W P(w) prior likelihood p( x w ) Perception A decision d ( x ) Action X Fig. 1 Goal: select d : X A Origin of SDT: WW2 radar operator Are the blobs enemy aircraft? Or just noise (e.g. clouds)? Decision has consequences: If you miss an aircraft, people might get killed If you mistake noise for aircraft, fuel, manpower & resources are wasted Radar screen 3
4 Decision outcomes SIGNAL: are the blobs real enemy aircraft? S S DECISION: should you alert the air force? Y N Hit Miss False alarm Correct reject SDT p X S p[ X S] probability noise aircraft X Blob size (stimulus intensity) SDT S S probability noise aircraft X X Blob size (stimulus intensity) 4
5 How should we set criterion? W A X = = { S, S} { Y, N} (, ) = π ( S), π ( S) [ ] prior p X S, p X S likelihood Y N S V V YS NS S VYS V NS gain How should we set criterion? ( ) = YS [ ] YS ( ) = NS [ ] + ( ) ( ) EBG Y V p S X V p S X EBG N V p S X V p S X NS Y EBG Y > EBG N [ ] ps X VYS + V > p S X V + V YS NS NS How should we set criterion? Y posterior odds [ ] ps X VYS + V > p S X V + V YS NS NS Y likelihood ratio prior odds [ ] π ( S) p X S VYS + V NS > p X S π ( S) V + V YS NS 5
6 How should we set criterion? Y Y [ ] π ( S) p X S VYS + V NS log > log + log p X S π ( S) V + V YS NS p[ X S] log > log β p X S log likelihood ratio Compare the LLR to a criterion, log β How should we set criterion? Y log p X S p X S > logβ log likelihood ratio Compare the LLR to a criterion, log β log p X S p X S > logβ p X S p X S > β S S probability p X S p X S noise aircraft X Blob size (stimulus intensity) 6
7 Suppose we vary prior odds but set the value term to 0. ( ) ( ) Y log p X S > log π S p X S π S Y p X S > π ( S ) p X S π ( S) + log V +V YS NS V YS +V NS Y p X S p X S S > π S π S S ( ) ( ) probability p X S p X S noise aircraft X Blob size (stimulus intensity) Y p X S p X S S ( ) ( ) > π S π S S We can fit data to estimate β We can compare to optimal β probability p X S p X S noise aircraft X Blob size (stimulus intensity) 7
8 Is the brain Bayesian? 4 Vary prior odds Estimate criterion b R 2 : logβˆ 0 1:1 9:1-2 1:9 γ : 0.36 p 0 : logβ Tanner, Swets, & Green (1956) Conservatism 4 R 2 : logβˆ 0 1:1 9:1-2 1:9 γ : 0.36 p 0 : logβ Tanner, Swets, & Green (1956) 8
9 Themes Managing uncertainty to maximize gain is the central task of a biological organism. The use of explicit cost and rewards allow us to probe a much wider range of behavior than previously explored (Trommershäuser et al, 2008) Movement planning is counterfactual reasoning. We can test SDT/BDT as a framework for modeling perception and action (Maloney & Mamassian, 2009) Aside The gain function is the organism s link to the environment. It represents a problem, posed by the environment, a problem that can rapidly change. A G( a, w ) Consequence gain W decision d ( x ) likelihood L ( w x ) X Only the luckiest organism can choose its gain function. Bayesian Decision Theory and the Planning of Movements L 9
10 Experimental Task Start of trial: display of fixation cross (1.5 s) L Experimental Task Display of response area, 500 ms before target onset (114.2 mm x 80.6 mm) L 10
11 Experimental Task Target display (700 ms) L Experimental Task L Experimental Task 100 L The green target is hit: +100 points
12 Experimental Task L -500 Experimental Task L The red target is hit: -500 points -500 Experimental Task L 12
13 Experimental Task L Scores add if both targets are hit: Experimental Task L Experimental Task You are too slow: -700 The screen is hit later than 700 ms after target display: -700 points. If you are on time but Miss the targets, 0. 13
14 Experimental Task Current score: 500 End of trial Choice among Movement Strategies mm 0 0 What should Paulina do? y hit -y mean (mm) Distribution of movement end points x hit -x mean (mm) Subject S4, = 3.62 mm, 72x15 = 1080 end points Expected Normal Value Q-Q Plot x hit -x mean (mm) y hit -y mean (mm) Observed Value 14
15 If there were no red penalty circle. Aim for center Select perceptual-motor strategies that minimize variance Harris & Wolpert (1998) Choice among Movement Strategies mm 0 0 What should Paulina do? Thought Experiment : -500 : 100 points (2.5 ) 100 points = 4.83 mm 15
16 Thought Experiment : -500 : 100 points (2.5 ) 100 points 100 points 200 points = 4.83 mm Thought Experiment : -500 : 100 points (2.5 ) 100 points 100 points 100 points 300 points = 4.83 mm Thought Experiment : -500 : 100 points (2.5 ) 100 points 100 points 100 points -400 points -100 points = 4.83 mm 16
17 Thought Experiment : -500 : 100 points (2.5 ) 100 points 100 points 100 points -400 points pts. per trial = 4.83 mm Thought Experiment : -500 : 100 points (2.5 ) pts. per trial 30.7 pts. per trial = 4.83 mm Thought Experiment : -500 : 100 points (2.5 ) pts. per trial 30.7 pts. per trial 25.5 pts. per trial = 4.83 mm 17
18 Thought Experiment : -500 : 100 points (2.5 ) pts. per trial 30.7 pts. per trial 25.5 pts. per trial 22.6 pts. per trial = 4.83 mm Expected value as function of mean movement end point (x,y): = 4.83 mm points per trial <-30 target: 100 penalty: -500 points per trial Thought Experiment penalty: 0 penalty: 100 penalty: 500 y x y x y x <-60 x, y: mean movement end point [mm] y [mm] y [mm] y [mm] = 4.83 mm x [mm] x [mm] x [mm] 18
19 Movement plans as lotteries (p 1, -500; p 2, -400; p 3, 100; p 4, 0) Movement plans as lotteries = 4 mm Lottery: (1.3%, -500; 30.3%, -400; 60.9%, 100; 7.5%, 0) Movement plans as lotteries Optimal aim point: lottery with MEV = 4 mm (6.6%, -500; 52.3%, -400; 37.0%, 100; 4.0%, 0) (1.3%, -500; 30.3%, -400; 60.9%, 100; 7.5%, 0) ( 0%, -500; 4.6%, -400; 62.6%, 100; 32.8%, 0) ( 0%, -500; 0.7%, -400; 37.6%, 100; 61.7%, 0) 19
20 Experiment 1 Reaching with Asymmetric Gain/Loss Julia Trommershäuser Trommershäuser, Maloney, Landy (2003) JOSA A Test of the model: Experiment 1 4 stimulus configurations: (varied within block) 2 penalty conditions: 0 and -500 points (varied between blocks) 5 practiced movers 1 session of data collection: 360 trials 24 data points per condition R = 9 mm General Methods: Training For all experiments: All subjects practice the task for 360 trials or more until their variance stabilizes. The timeout limit is gradually decreased to 700 ms during training. There are no penalties during training (the concept is never mentioned). We verify that each subject s movement variance has stabilized. They are told only to (0) +100 make money. 20
21 General Methods: Training For all experiments: All subjects practice the task for 360 trials or more until their variance stabilizes. The timeout limit is gradually decreased to 700 ms during training. There are no penalties during training (the concept is never mentioned). We verify that each subject s movement variance has stabilized. They are told only to (0) +100 make money. Model prediction: Results: Experiment 1 model, penalty = 0 Subject S5, = 2.99 mm Results: Experiment 1 Model prediction: configuration 1 model, penalty = 0 x model, penalty = 500 Subject S5, = 2.99 mm 21
22 Results: Experiment 1 Model prediction: configuration 2 model, penalty = 0 x model, penalty = 500 Subject S5, = 2.99 mm Results: Experiment 1 Model prediction: configuration 3 model, penalty = 0 x model, penalty = 500 Subject S5, = 2.99 mm Results: Experiment 1 Model prediction: configuration 4 model, penalty = 0 x model, penalty = 500 Subject S5, = 2.99 mm 22
23 Results: Experiment 1 Comparison with experiment exp., penalty = 0 exp., penalty = 500 x model, penalty = 500 Subject S5, = 2.99 mm Results: Experiment 1 S1 S2 S3 S4 S5 exp., penalty = 0 exp., penalty = 500 x model, penalty = 500 Results: Experiment 1 S % S % S % S % S % exp., penalty = 0 exp., penalty = 500 x model, penalty =
24 Two tasks 1. Movement task: Aim for a point on the screen Extensive training 2. Decision task: Select which point to aim for No explicit training Learning? Hill-Climbing? mean end point offset (mm) Hypothetical Trend trial number 24
25 Conclusions The use of explicit cost and rewards allow us to probe a much wider range of behavior than previously explored. Managing uncertainty to maximize gain is the central task of a biological organism. 25
Statistical Decision Theory II: Costs, Rewards, & Ideal Actors
Statistical Decision Theory II: Costs, Rewards, & Ideal Actors Loss Functions Statistical decision theory involves determining which decisions (or actions) are best It formalizes good versus bad outcomes
More informationConcerns of the Psychophysicist. Three methods for measuring perception. Yes/no method of constant stimuli. Detection / discrimination.
Three methods for measuring perception Concerns of the Psychophysicist. Magnitude estimation 2. Matching 3. Detection/discrimination Bias/ Attentiveness Strategy/Artifactual Cues History of stimulation
More informationSupplemental Information. Noise and Correlations. in Parallel Perceptual Decision Making. Thomas U. Otto and Pascal Mamassian
Current Biology, Volume 22 Supplemental Information Noise and Correlations in Parallel Perceptual Decision Making Thomas U. Otto and Pascal Mamassian Supplemental Inventory Figure S1 (related to Figure
More informationSupplementary Figure 1. Characterization of the single-photon quantum light source based on spontaneous parametric down-conversion (SPDC).
.2 Classical light source.8 g (2) ().6.4.2 EMCCD SPAD 2 3.2.4.6.8..2.4.6.8.2 Mean number of photon pairs per pump pulse 4 5 6 7 8 9 2 3 4 Supplementary Figure. Characterization of the single-photon quantum
More informationBayesian Concept Learning
Learning from positive and negative examples Bayesian Concept Learning Chen Yu Indiana University With both positive and negative examples, it is easy to define a boundary to separate these two. Just with
More information(1) Introduction to Bayesian statistics
Spring, 2018 A motivating example Student 1 will write down a number and then flip a coin If the flip is heads, they will honestly tell student 2 if the number is even or odd If the flip is tails, they
More informationModel Averaging (Bayesian Learning)
Model Averaging (Bayesian Learning) We want to predict the output Y of a new case that has input X = x given the training examples e: p(y x e) = m M P(Y m x e) = m M P(Y m x e)p(m x e) = m M P(Y m x)p(m
More informationSupplemental Data: Appendix
Supplemental Data: Appendix A Demonstration of model identifiability A.1 Identifiability of of 2-ADC model In this section, we demonstrate that there is a one-to-one mapping from the set of parameters
More informationDetection theory. H 0 : x[n] = w[n]
Detection Theory Detection theory A the last topic of the course, we will briefly consider detection theory. The methods are based on estimation theory and attempt to answer questions such as Is a signal
More informationLecture 4: Types of errors. Bayesian regression models. Logistic regression
Lecture 4: Types of errors. Bayesian regression models. Logistic regression A Bayesian interpretation of regularization Bayesian vs maximum likelihood fitting more generally COMP-652 and ECSE-68, Lecture
More informationA DECISION-MAKING THEORY OF VISUAL DETECTION l
Piycholottcal Review Vol 61, No 6. 1954 A DECISION-MAKING THEORY OF VISUAL DETECTION l WILSON P. TANNER, JR. AND JOHN A. SWETS This paper is concerned with the human observer's behavior in detecting light
More information+ + ( + ) = Linear recurrent networks. Simpler, much more amenable to analytic treatment E.g. by choosing
Linear recurrent networks Simpler, much more amenable to analytic treatment E.g. by choosing + ( + ) = Firing rates can be negative Approximates dynamics around fixed point Approximation often reasonable
More informationMathematical Tools for Neuroscience (NEU 314) Princeton University, Spring 2016 Jonathan Pillow. Homework 8: Logistic Regression & Information Theory
Mathematical Tools for Neuroscience (NEU 34) Princeton University, Spring 206 Jonathan Pillow Homework 8: Logistic Regression & Information Theory Due: Tuesday, April 26, 9:59am Optimization Toolbox One
More informationBayesian Information Criterion as a Practical Alternative to Null-Hypothesis Testing Michael E. J. Masson University of Victoria
Bayesian Information Criterion as a Practical Alternative to Null-Hypothesis Testing Michael E. J. Masson University of Victoria Presented at the annual meeting of the Canadian Society for Brain, Behaviour,
More informationThe Naïve Bayes Classifier. Machine Learning Fall 2017
The Naïve Bayes Classifier Machine Learning Fall 2017 1 Today s lecture The naïve Bayes Classifier Learning the naïve Bayes Classifier Practical concerns 2 Today s lecture The naïve Bayes Classifier Learning
More informationBayesian Inference. Will Penny. 24th February Bayesian Inference. Will Penny. Bayesian Inference. References
24th February 2011 Given probabilities p(a), p(b), and the joint probability p(a, B), we can write the conditional probabilities p(b A) = p(a B) = p(a, B) p(a) p(a, B) p(b) Eliminating p(a, B) gives p(b
More informationEncoding or decoding
Encoding or decoding Decoding How well can we learn what the stimulus is by looking at the neural responses? We will discuss two approaches: devise and evaluate explicit algorithms for extracting a stimulus
More informationSequential Decisions
Sequential Decisions A Basic Theorem of (Bayesian) Expected Utility Theory: If you can postpone a terminal decision in order to observe, cost free, an experiment whose outcome might change your terminal
More informationSequential Decisions
Sequential Decisions A Basic Theorem of (Bayesian) Expected Utility Theory: If you can postpone a terminal decision in order to observe, cost free, an experiment whose outcome might change your terminal
More information1 Descriptive statistics. 2 Scores and probability distributions. 3 Hypothesis testing and one-sample t-test. 4 More on t-tests
Overall Overview INFOWO Statistics lecture S3: Hypothesis testing Peter de Waal Department of Information and Computing Sciences Faculty of Science, Universiteit Utrecht 1 Descriptive statistics 2 Scores
More informationA Framework for Testing Identifiability of Bayesian Models of Perception
A Framework for Testing Identifiability of Bayesian Models of Perception Luigi Acerbi,2 Wei Ji Ma 2 Sethu Vijayakumar School of Informatics, University of Edinburgh, UK 2 Center for Neural Science & Department
More informationMachine Learning, Midterm Exam: Spring 2009 SOLUTION
10-601 Machine Learning, Midterm Exam: Spring 2009 SOLUTION March 4, 2009 Please put your name at the top of the table below. If you need more room to work out your answer to a question, use the back of
More informationLecture 5: Likelihood ratio tests, Neyman-Pearson detectors, ROC curves, and sufficient statistics. 1 Executive summary
ECE 830 Spring 207 Instructor: R. Willett Lecture 5: Likelihood ratio tests, Neyman-Pearson detectors, ROC curves, and sufficient statistics Executive summary In the last lecture we saw that the likelihood
More informationMultivariate statistical methods and data mining in particle physics
Multivariate statistical methods and data mining in particle physics RHUL Physics www.pp.rhul.ac.uk/~cowan Academic Training Lectures CERN 16 19 June, 2008 1 Outline Statement of the problem Some general
More informationIntroductory Econometrics
Session 4 - Testing hypotheses Roland Sciences Po July 2011 Motivation After estimation, delivering information involves testing hypotheses Did this drug had any effect on the survival rate? Is this drug
More information!) + log(t) # n i. The last two terms on the right hand side (RHS) are clearly independent of θ and can be
Supplementary Materials General case: computing log likelihood We first describe the general case of computing the log likelihood of a sensory parameter θ that is encoded by the activity of neurons. Each
More informationLecture 3. STAT161/261 Introduction to Pattern Recognition and Machine Learning Spring 2018 Prof. Allie Fletcher
Lecture 3 STAT161/261 Introduction to Pattern Recognition and Machine Learning Spring 2018 Prof. Allie Fletcher Previous lectures What is machine learning? Objectives of machine learning Supervised and
More informationPopulation Coding. Maneesh Sahani Gatsby Computational Neuroscience Unit University College London
Population Coding Maneesh Sahani maneesh@gatsby.ucl.ac.uk Gatsby Computational Neuroscience Unit University College London Term 1, Autumn 2010 Coding so far... Time-series for both spikes and stimuli Empirical
More informationMATH 446/546 Homework 2: Due October 8th, 2014
MATH 446/546 Homework 2: Due October 8th, 2014 Answer the following questions. Some of which come from Winston s text book. 1. We are going to invest $1,000 for a period of 6 months. Two potential investments
More informationHypothesis testing (cont d)
Hypothesis testing (cont d) Ulrich Heintz Brown University 4/12/2016 Ulrich Heintz - PHYS 1560 Lecture 11 1 Hypothesis testing Is our hypothesis about the fundamental physics correct? We will not be able
More informationPerformance Evaluation and Comparison
Outline Hong Chang Institute of Computing Technology, Chinese Academy of Sciences Machine Learning Methods (Fall 2012) Outline Outline I 1 Introduction 2 Cross Validation and Resampling 3 Interval Estimation
More informationRecipes for the Linear Analysis of EEG and applications
Recipes for the Linear Analysis of EEG and applications Paul Sajda Department of Biomedical Engineering Columbia University Can we read the brain non-invasively and in real-time? decoder 1001110 if YES
More informationPower Analysis. Introduction to Power
Power Analysis Dr. J. Kyle Roberts Southern Methodist University Simmons School of Education and Human Development Department of Teaching and Learning When testing a specific null hypothesis (H 0 ), we
More informationIntroduction to Bayesian Learning. Machine Learning Fall 2018
Introduction to Bayesian Learning Machine Learning Fall 2018 1 What we have seen so far What does it mean to learn? Mistake-driven learning Learning by counting (and bounding) number of mistakes PAC learnability
More informationLinear classifiers Lecture 3
Linear classifiers Lecture 3 David Sontag New York University Slides adapted from Luke Zettlemoyer, Vibhav Gogate, and Carlos Guestrin ML Methodology Data: labeled instances, e.g. emails marked spam/ham
More informationTesting Simple Hypotheses R.L. Wolpert Institute of Statistics and Decision Sciences Duke University, Box Durham, NC 27708, USA
Testing Simple Hypotheses R.L. Wolpert Institute of Statistics and Decision Sciences Duke University, Box 90251 Durham, NC 27708, USA Summary: Pre-experimental Frequentist error probabilities do not summarize
More informationPATTERN RECOGNITION AND MACHINE LEARNING
PATTERN RECOGNITION AND MACHINE LEARNING Chapter 1. Introduction Shuai Huang April 21, 2014 Outline 1 What is Machine Learning? 2 Curve Fitting 3 Probability Theory 4 Model Selection 5 The curse of dimensionality
More information2. What are the tradeoffs among different measures of error (e.g. probability of false alarm, probability of miss, etc.)?
ECE 830 / CS 76 Spring 06 Instructors: R. Willett & R. Nowak Lecture 3: Likelihood ratio tests, Neyman-Pearson detectors, ROC curves, and sufficient statistics Executive summary In the last lecture we
More informationIntelligent Systems Statistical Machine Learning
Intelligent Systems Statistical Machine Learning Carsten Rother, Dmitrij Schlesinger WS2014/2015, Our tasks (recap) The model: two variables are usually present: - the first one is typically discrete k
More informationWays to make neural networks generalize better
Ways to make neural networks generalize better Seminar in Deep Learning University of Tartu 04 / 10 / 2014 Pihel Saatmann Topics Overview of ways to improve generalization Limiting the size of the weights
More informationThe Bayesian Brain. Robert Jacobs Department of Brain & Cognitive Sciences University of Rochester. May 11, 2017
The Bayesian Brain Robert Jacobs Department of Brain & Cognitive Sciences University of Rochester May 11, 2017 Bayesian Brain How do neurons represent the states of the world? How do neurons represent
More informationBayesian Gaussian / Linear Models. Read Sections and 3.3 in the text by Bishop
Bayesian Gaussian / Linear Models Read Sections 2.3.3 and 3.3 in the text by Bishop Multivariate Gaussian Model with Multivariate Gaussian Prior Suppose we model the observed vector b as having a multivariate
More informationBayesian Decision Theory
Bayesian Decision Theory Selim Aksoy Department of Computer Engineering Bilkent University saksoy@cs.bilkent.edu.tr CS 551, Fall 2017 CS 551, Fall 2017 c 2017, Selim Aksoy (Bilkent University) 1 / 46 Bayesian
More informationHUMAN REPRESENTATION OF VISUO-MOTOR UNCERTAINTY
HUMAN REPRESENTATION OF VISUO-MOTOR UNCERTAINTY AS MIXTURES OF ORTHOGONAL BASIS DISTRIBUTIONS HANG ZHANG 1,2,3,4,5,6 NATHANIEL D. DAW 4,5,6 LAURENCE T. MALONEY 4,5,6 1. Department of Psychology and Beijing
More informationPATTERN RECOGNITION AND MACHINE LEARNING
PATTERN RECOGNITION AND MACHINE LEARNING Slide Set 3: Detection Theory January 2018 Heikki Huttunen heikki.huttunen@tut.fi Department of Signal Processing Tampere University of Technology Detection theory
More informationLogistic Regression. Machine Learning Fall 2018
Logistic Regression Machine Learning Fall 2018 1 Where are e? We have seen the folloing ideas Linear models Learning as loss minimization Bayesian learning criteria (MAP and MLE estimation) The Naïve Bayes
More informationBayesian Learning (II)
Universität Potsdam Institut für Informatik Lehrstuhl Maschinelles Lernen Bayesian Learning (II) Niels Landwehr Overview Probabilities, expected values, variance Basic concepts of Bayesian learning MAP
More informationSRNDNA Model Fitting in RL Workshop
SRNDNA Model Fitting in RL Workshop yael@princeton.edu Topics: 1. trial-by-trial model fitting (morning; Yael) 2. model comparison (morning; Yael) 3. advanced topics (hierarchical fitting etc.) (afternoon;
More informationMark your answers ON THE EXAM ITSELF. If you are not sure of your answer you may wish to provide a brief explanation.
CS 189 Spring 2015 Introduction to Machine Learning Midterm You have 80 minutes for the exam. The exam is closed book, closed notes except your one-page crib sheet. No calculators or electronic items.
More informationINTRODUCTION TO PATTERN RECOGNITION
INTRODUCTION TO PATTERN RECOGNITION INSTRUCTOR: WEI DING 1 Pattern Recognition Automatic discovery of regularities in data through the use of computer algorithms With the use of these regularities to take
More informationECE521 Lecture7. Logistic Regression
ECE521 Lecture7 Logistic Regression Outline Review of decision theory Logistic regression A single neuron Multi-class classification 2 Outline Decision theory is conceptually easy and computationally hard
More informationBayesian Linear Regression [DRAFT - In Progress]
Bayesian Linear Regression [DRAFT - In Progress] David S. Rosenberg Abstract Here we develop some basics of Bayesian linear regression. Most of the calculations for this document come from the basic theory
More informationADVANCED MACHINE LEARNING ADVANCED MACHINE LEARNING. Non-linear regression techniques Part - II
1 Non-linear regression techniques Part - II Regression Algorithms in this Course Support Vector Machine Relevance Vector Machine Support vector regression Boosting random projections Relevance vector
More information3 Neural Decoding. 3.1 Encoding and Decoding. (r 1, r 2,..., r N ) for N neurons is a list of spike-count firing rates, although,
3 Neural Decoding 3.1 Encoding and Decoding In chapters 1 and 2, we considered the problem of predicting neural responses to known stimuli. The nervous system faces the reverse problem, determining what
More informationECLT 5810 Linear Regression and Logistic Regression for Classification. Prof. Wai Lam
ECLT 5810 Linear Regression and Logistic Regression for Classification Prof. Wai Lam Linear Regression Models Least Squares Input vectors is an attribute / feature / predictor (independent variable) The
More informationCOMP3702/7702 Artificial Intelligence Lecture 11: Introduction to Machine Learning and Reinforcement Learning. Hanna Kurniawati
COMP3702/7702 Artificial Intelligence Lecture 11: Introduction to Machine Learning and Reinforcement Learning Hanna Kurniawati Today } What is machine learning? } Where is it used? } Types of machine learning
More informationBayesian Decision Theory
Introduction to Pattern Recognition [ Part 4 ] Mahdi Vasighi Remarks It is quite common to assume that the data in each class are adequately described by a Gaussian distribution. Bayesian classifier is
More informationIntelligent Systems Statistical Machine Learning
Intelligent Systems Statistical Machine Learning Carsten Rother, Dmitrij Schlesinger WS2015/2016, Our model and tasks The model: two variables are usually present: - the first one is typically discrete
More informationCourse 16:198:520: Introduction To Artificial Intelligence Lecture 13. Decision Making. Abdeslam Boularias. Wednesday, December 7, 2016
Course 16:198:520: Introduction To Artificial Intelligence Lecture 13 Decision Making Abdeslam Boularias Wednesday, December 7, 2016 1 / 45 Overview We consider probabilistic temporal models where the
More informationMachine Learning Ensemble Learning I Hamid R. Rabiee Jafar Muhammadi, Alireza Ghasemi Spring /
Machine Learning Ensemble Learning I Hamid R. Rabiee Jafar Muhammadi, Alireza Ghasemi Spring 2015 http://ce.sharif.edu/courses/93-94/2/ce717-1 / Agenda Combining Classifiers Empirical view Theoretical
More informationMachine Learning, Fall 2009: Midterm
10-601 Machine Learning, Fall 009: Midterm Monday, November nd hours 1. Personal info: Name: Andrew account: E-mail address:. You are permitted two pages of notes and a calculator. Please turn off all
More informationAdvanced statistical methods for data analysis Lecture 1
Advanced statistical methods for data analysis Lecture 1 RHUL Physics www.pp.rhul.ac.uk/~cowan Universität Mainz Klausurtagung des GK Eichtheorien exp. Tests... Bullay/Mosel 15 17 September, 2008 1 Outline
More informationReview of Probability, Expected Utility
Review of Probability, Expected Utility Economics 302 - Microeconomic Theory II: Strategic Behavior Instructor: Songzi Du compiled by Shih En Lu Simon Fraser University May 8, 2018 ECON 302 (SFU) Lecture
More informationLeast Squares Regression
CIS 50: Machine Learning Spring 08: Lecture 4 Least Squares Regression Lecturer: Shivani Agarwal Disclaimer: These notes are designed to be a supplement to the lecture. They may or may not cover all the
More informationToday. Calculus. Linear Regression. Lagrange Multipliers
Today Calculus Lagrange Multipliers Linear Regression 1 Optimization with constraints What if I want to constrain the parameters of the model. The mean is less than 10 Find the best likelihood, subject
More informationBios 6649: Clinical Trials - Statistical Design and Monitoring
Bios 6649: Clinical Trials - Statistical Design and Monitoring Spring Semester 2015 John M. Kittelson Department of Biostatistics & nformatics Colorado School of Public Health University of Colorado Denver
More informationCOMP 551 Applied Machine Learning Lecture 21: Bayesian optimisation
COMP 55 Applied Machine Learning Lecture 2: Bayesian optimisation Associate Instructor: (herke.vanhoof@mcgill.ca) Class web page: www.cs.mcgill.ca/~jpineau/comp55 Unless otherwise noted, all material posted
More informationMachine Learning. Lecture 4: Regularization and Bayesian Statistics. Feng Li. https://funglee.github.io
Machine Learning Lecture 4: Regularization and Bayesian Statistics Feng Li fli@sdu.edu.cn https://funglee.github.io School of Computer Science and Technology Shandong University Fall 207 Overfitting Problem
More informationCIS 390 Fall 2016 Robotics: Planning and Perception Final Review Questions
CIS 390 Fall 2016 Robotics: Planning and Perception Final Review Questions December 14, 2016 Questions Throughout the following questions we will assume that x t is the state vector at time t, z t is the
More informationIntroduction to Detection Theory
Introduction to Detection Theory Detection Theory (a.k.a. decision theory or hypothesis testing) is concerned with situations where we need to make a decision on whether an event (out of M possible events)
More informationClassifier performance evaluation
Classifier performance evaluation Václav Hlaváč Czech Technical University in Prague Czech Institute of Informatics, Robotics and Cybernetics 166 36 Prague 6, Jugoslávských partyzánu 1580/3, Czech Republic
More informationEfficient coding of natural images with a population of noisy Linear-Nonlinear neurons
Efficient coding of natural images with a population of noisy Linear-Nonlinear neurons Yan Karklin and Eero P. Simoncelli NYU Overview Efficient coding is a well-known objective for the evaluation and
More informationNeural Decoding. Chapter Encoding and Decoding
Chapter 3 Neural Decoding 3.1 Encoding and Decoding In chapters 1 and 2, we considered the problem of predicting neural responses to known stimuli. The nervous system faces the reverse problem, determining
More informationBayesian Updating: Discrete Priors: Spring
Bayesian Updating: Discrete Priors: 18.05 Spring 2017 http://xkcd.com/1236/ Learning from experience Which treatment would you choose? 1. Treatment 1: cured 100% of patients in a trial. 2. Treatment 2:
More informationLECTURE NOTE #3 PROF. ALAN YUILLE
LECTURE NOTE #3 PROF. ALAN YUILLE 1. Three Topics (1) Precision and Recall Curves. Receiver Operating Characteristic Curves (ROC). What to do if we do not fix the loss function? (2) The Curse of Dimensionality.
More informationStatistical Inference
Statistical Inference Classical and Bayesian Methods Class 6 AMS-UCSC Thu 26, 2012 Winter 2012. Session 1 (Class 6) AMS-132/206 Thu 26, 2012 1 / 15 Topics Topics We will talk about... 1 Hypothesis testing
More informationσ(a) = a N (x; 0, 1 2 ) dx. σ(a) = Φ(a) =
Until now we have always worked with likelihoods and prior distributions that were conjugate to each other, allowing the computation of the posterior distribution to be done in closed form. Unfortunately,
More informationBioinformatics: Network Analysis
Bioinformatics: Network Analysis Model Fitting COMP 572 (BIOS 572 / BIOE 564) - Fall 2013 Luay Nakhleh, Rice University 1 Outline Parameter estimation Model selection 2 Parameter Estimation 3 Generally
More informationNatural Language Processing. Classification. Features. Some Definitions. Classification. Feature Vectors. Classification I. Dan Klein UC Berkeley
Natural Language Processing Classification Classification I Dan Klein UC Berkeley Classification Automatically make a decision about inputs Example: document category Example: image of digit digit Example:
More informationCS 446 Machine Learning Fall 2016 Nov 01, Bayesian Learning
CS 446 Machine Learning Fall 206 Nov 0, 206 Bayesian Learning Professor: Dan Roth Scribe: Ben Zhou, C. Cervantes Overview Bayesian Learning Naive Bayes Logistic Regression Bayesian Learning So far, we
More informationComputational Cognitive Science
Computational Cognitive Science Lecture 9: A Bayesian model of concept learning Chris Lucas School of Informatics University of Edinburgh October 16, 218 Reading Rules and Similarity in Concept Learning
More informationStat Lecture 20. Last class we introduced the covariance and correlation between two jointly distributed random variables.
Stat 260 - Lecture 20 Recap of Last Class Last class we introduced the covariance and correlation between two jointly distributed random variables. Today: We will introduce the idea of a statistic and
More informationJoint Factor Analysis for Speaker Verification
Joint Factor Analysis for Speaker Verification Mengke HU ASPITRG Group, ECE Department Drexel University mengke.hu@gmail.com October 12, 2012 1/37 Outline 1 Speaker Verification Baseline System Session
More informationUniversität Potsdam Institut für Informatik Lehrstuhl Maschinelles Lernen. Bayesian Learning. Tobias Scheffer, Niels Landwehr
Universität Potsdam Institut für Informatik Lehrstuhl Maschinelles Lernen Bayesian Learning Tobias Scheffer, Niels Landwehr Remember: Normal Distribution Distribution over x. Density function with parameters
More informationCheng Soon Ong & Christian Walder. Canberra February June 2018
Cheng Soon Ong & Christian Walder Research Group and College of Engineering and Computer Science Canberra February June 2018 (Many figures from C. M. Bishop, "Pattern Recognition and ") 1of 143 Part IV
More informationData Analysis for an Absolute Identification Experiment. Randomization with Replacement. Randomization without Replacement
Data Analysis for an Absolute Identification Experiment 1 Randomization with Replacement Imagine that you have k containers for the k stimulus alternatives The i th container has a fixed number of copies
More informationHypothesis Testing. Part I. James J. Heckman University of Chicago. Econ 312 This draft, April 20, 2006
Hypothesis Testing Part I James J. Heckman University of Chicago Econ 312 This draft, April 20, 2006 1 1 A Brief Review of Hypothesis Testing and Its Uses values and pure significance tests (R.A. Fisher)
More informationLecture 2 Machine Learning Review
Lecture 2 Machine Learning Review CMSC 35246: Deep Learning Shubhendu Trivedi & Risi Kondor University of Chicago March 29, 2017 Things we will look at today Formal Setup for Supervised Learning Things
More informationSearch for Inspiral GW Signals Associated with Short GRBs using Artificial Neural Networks
Search for Inspiral GW Signals Associated with Short GRBs using Artificial Neural Networks Kyungmin Kim and Hyun Kyu Lee Hanyang Univ. In collaboration with Y.-M. Kim, C.-H. Lee (PNU), J.J. Oh, S.H. Oh,
More informationStatistical Methods in Particle Physics
Statistical Methods in Particle Physics Lecture 11 January 7, 2013 Silvia Masciocchi, GSI Darmstadt s.masciocchi@gsi.de Winter Semester 2012 / 13 Outline How to communicate the statistical uncertainty
More informationCSC321 Lecture 18: Learning Probabilistic Models
CSC321 Lecture 18: Learning Probabilistic Models Roger Grosse Roger Grosse CSC321 Lecture 18: Learning Probabilistic Models 1 / 25 Overview So far in this course: mainly supervised learning Language modeling
More informationLecture 3. Linear Regression II Bastian Leibe RWTH Aachen
Advanced Machine Learning Lecture 3 Linear Regression II 02.11.2015 Bastian Leibe RWTH Aachen http://www.vision.rwth-aachen.de/ leibe@vision.rwth-aachen.de This Lecture: Advanced Machine Learning Regression
More informationNon-Bayesian Classifiers Part II: Linear Discriminants and Support Vector Machines
Non-Bayesian Classifiers Part II: Linear Discriminants and Support Vector Machines Selim Aksoy Department of Computer Engineering Bilkent University saksoy@cs.bilkent.edu.tr CS 551, Fall 2018 CS 551, Fall
More informationDetection and Estimation Chapter 1. Hypothesis Testing
Detection and Estimation Chapter 1. Hypothesis Testing Husheng Li Min Kao Department of Electrical Engineering and Computer Science University of Tennessee, Knoxville Spring, 2015 1/20 Syllabus Homework:
More informationHomework 2 Solutions Kernel SVM and Perceptron
Homework 2 Solutions Kernel SVM and Perceptron CMU 1-71: Machine Learning (Fall 21) https://piazza.com/cmu/fall21/17115781/home OUT: Sept 25, 21 DUE: Oct 8, 11:59 PM Problem 1: SVM decision boundaries
More informationThe Diffusion Model of Speeded Choice, from a Rational Perspective
The Diffusion Model of Speeded Choice, from a Rational Perspective Matt Jones, University of Colorado August 7, 017 1 Binary Decision Tasks This chapter considers tasks in which an experimental subject
More informationSTA414/2104 Statistical Methods for Machine Learning II
STA414/2104 Statistical Methods for Machine Learning II Murat A. Erdogdu & David Duvenaud Department of Computer Science Department of Statistical Sciences Lecture 3 Slide credits: Russ Salakhutdinov Announcements
More informationMachine Learning 4771
Machine Learning 4771 Instructor: Tony Jebara Topic 11 Maximum Likelihood as Bayesian Inference Maximum A Posteriori Bayesian Gaussian Estimation Why Maximum Likelihood? So far, assumed max (log) likelihood
More information6.867 Machine Learning
6.867 Machine Learning Problem Set 2 Due date: Wednesday October 6 Please address all questions and comments about this problem set to 6867-staff@csail.mit.edu. You will need to use MATLAB for some of
More informationAchievable rates for pattern recognition
Achievable rates for pattern recognition M Brandon Westover Joseph A O Sullivan Washington University in Saint Louis Departments of Physics and Electrical Engineering Goals of Information Theory Models
More information