MEDICAL IMAGING INFORMATICS: Lecture # 1
|
|
- Marcus O’Brien’
- 5 years ago
- Views:
Transcription
1 MEDICAL IMAGIG IFORMATICS: Lecture # 1 Basics of Medical Imaging Informatics: Estimation Theory orbert Schuff Professor of Radiology Medical Center and UCSF orbert.schuff@ucsf.edu Medical Imaging Informatics 2010, schuff Slide 1/31
2 What Is Medical Imaging Informatics? Signal Processing Digital Image Acquisition Image Processing and Enhancement Data Mining Computational anatomy Statistics Databases Data-mining Workflow and Process Modeling and Simulation Data Management Picture Archiving and Communication System (PACS) Imaging Informatics for the Enterprise Image-Enabled Electronic Medical Records Radiology Information Systems (RIS) and Hospital Information Systems (HIS) Quality Assurance Archive Integrity and Security Data Visualization Image Data Compression 3D, Visualization and Multi-media DICOM, HL7 and other Standards Teleradiology Imaging Vocabularies and Ontologies Transforming the Radiological Interpretation Process (TRIP)[2] Computer-Aided Detection and Diagnosis (CAD). Radiology Informatics Education Etc. Slide 2/31
3 What Is The Focus Of This Course? Learn using computational tools to maximize information and knowledge gain Pro-active Improve Data collection Refine Model Measurements Image Model knowledge Re-active Extract information Compare with model Slide 3/31
4 Challenge: Maximize Information Gain 1. Q: How can we estimate quantities of interest from a given set of uncertain (noise) measurements? A: Apply estimation theory (1 st lecture by orbert) 2. Q: How can we measure (quantify) information? A: Apply information theory (2 nd lecture by Wang) Slide 4/31
5 Estimation Theory: Motivation Example I Gray/White Matter Segmentation Hypothetical Histogram Intensity GM/WM overlap 50:50; Can we do better than flipping a coin? Slide 5/31
6 Estimation Theory: Motivation Example II Goal: Capture dynamic signal on a static background High signal to noise Poor signal to noise D. Feinberg Advanced MRI Technologies, Sebastopol, CA Slide 6/31
7 Estimation Theory: Motivation Example III Diffusion Spectrum Imaging Human Cingulum Bundle Goal: Capture directions of fiber bundles Dr. Van Wedeen, MGH Slide 7/31
8 Basic Concepts of Modeling Θ: target of interest and unknown ρ: measurement Θ ) : Estimator - a good guess of Θ based on measurements Cartoon adapted from: Rajesh P.. Rao, Bruno A. Olshausen Probabilistic Models of the Brain. MIT Press Slide 8/31
9 Deterministic Model ϕ = Hθ + M noise = number of measurements M = number of states, M=1 is possible Usually > M and noise 2 > 0 The model is deterministic, because discrete values of Θ are solutions. ote: 1) we make no assumption about Θ 2) Each value is as likely as any another value What is the best estimator under these circumstances? Slide 9/31
10 Least-Squares Estimator (LSE) The best what we can do is minimizing noise: Minimizing E LSE with regard to θ leads to E LSE ϕ ˆ Hθ LSE = 0 T Hϕ θˆ = 0 = 1 min 2 ( T HH) noise θ ) LSE LSE = ( T ) HH 1 LSE is popular choice for model fitting Useful for obtaining a descriptive measure But LSE makes no assumptions about distributions of data or parameters Has no basis for statistics deterministic model 2 H T ϕ n Slide 10/31
11 Prominent Examples of LSE 1 = Mean Value: ˆ θ ϕ( j) mean j= 1 1 θvar iance = ϕ θ 1 Variance ˆ ( j) j= 1 ( ˆ ) 2 mean Amplitude: Frequency: Phase: Decay: θˆ 1 θˆ 2 θˆ 3 θˆ 4 Slide 11/31
12 Likelihood Model Likelihood L ( Φ ) = p( ϕ Φ) ϕ Pretend we know something about Θ We perform measurements for all possible values of Θ We obtain the likelihood function of Θ given our measurements ρ ote: Θ is random ϕ is a fixed parameter Likelihood is a function of both the unknown Θ and known ϕ Slide 12/31
13 Likelihood Model (cont d) L ϕ ( Φ ) = p( ϕ Φ) ew Goal: Find an estimator which gives the most likely probability distribution underlying Φ L ϕ ( ) Slide 13/31
14 Maximum Likelihood Estimator (MLE) Goal: Find estimator which gives the most likely probability distribution underlying x. Max likelihood function ) Φ MLE ( ) = max p ϕ Φ θ MLE can be found by d dφ ( ϕ ) ln p Φ = 0 θ= θ MLE Slide 14/31
15 Example I: MLE Of ormal Distribution ormal distribution 2 1 p( ϕ Φ, σ ) exp ϕ 2 ( j) 2 Φ σ j= 1 ( ) 2 log of the normal distribution (normd) 2 ( ϕ ) ( ( j) ) 2 Φ σ ϕ Φ 2 ln p, 1 2 σ j= 1 ormal Distribution Log ormal Distribution MLE of the mean (1 st derivative): d ) dφ MLE 1 ) ln p() = ( ϕ ( j) ΦMLE ) = 0 2 ˆ 2 σ j= 1 ) 1 Φ MLE = j= 1 ϕ ( j) Slide 15/31
16 Example II: MLE Of Binominal Distribution (Coin Toss) Distribution function f(y n,w): n= number of tosses w= probability of success f(y n=10,w=7) f(y n=10,w=3) Slide 16/31 y
17 MLE Of Coin Toss (cont d) Goal: Given the observed data f (y w=0.7, n=10), find the parameter Φ MLE that most likely produced the data. L( Φ y = 7, n= 10) MLE Slide 17/31
18 MLE Of Coin Toss (cont d) Likelihood function of coin tosses n! y L( Φ y) = Φ 1 Φ!! ( y )( n y) ( ) n y log likelihood function ( ) ln L w y = n! ln + yln w+ n y ln 1 w!! ( y )( n y) ( ) ( ) Slide 18/31
19 MLE Of Coin Toss Evaluate MLE equation (1 st derivative) () y ( ) dln L n y = = dφ ΦML 1 Φ MLE E MLE 0 y nφ MLE y = = 0 ΦMLE = Φ (1 Φ ) n MLE MLE According to the MLE principle, the distribution f(y/n) for a given n is the most likely distribution to have generated the observed data of y. Slide 19/31
20 Relationship between MLE and LSE Assume: Θ is independent of noise MLE and noise have the same distribution ( ϕ Φ ) = ( ϕ Φ Φ) p p noise θ H noise is zero mean and gaussian p(ρ Θ) is maximized when LSE is minimized Slide 20/31
21 Prior knowledge Bayesian Model ow, the daemon comes into play, but we know The daemon s preferences for Θ (prior knowledge). prior ( Φ ) = p ( Φ) ew Goal: Find the estimator which gives the most likely probability distribution of Θ given everything we know. Slide 21/31
22 Bayesian Model ( θ) = ( Φ) ( Φ) posterior C L p ϕ ϕ ϕ Slide 22/31
23 Maximum A-Posteriori (MAP) Estimator Goal: Find the most likely Θ MAP (max. posterior density of ) given ϕ. ˆ θ = max L ϕ θ p ϕ MAP ( ) ( ) Maximize joint density θ MAO can be found by d ln L( ϕ θ) p( θ) = ln p( ϕ θ) + ln p( θ) = 0 dθ θ θ Slide 23/31
24 Example III: MAP Of ormal Distribution We have random sample: ( 2) ( 2) 1 1 ln Lϕ ˆˆ, ˆˆ, ( ( ) ˆ ) ( ˆ μσ p μσ = ϕ j μϕ p μmle ) = 0 uˆ ˆ σ ˆ σ 2 2 ϕ j= 1 μ j= 1 The sample mean of MAP is: ) Φ MAP = σ 2 μ 2 2 ϕ + T μ j σ σ = 1 ϕ ( j) If we do not have prior information on μ, σ μ inf or T inf μˆ μˆ μˆ MAP ML, LSE Slide 24/31
25 Posterior Distribution and Decision Rules p(θ ρ) Slide 25/31 Θ MSE Θ MAP Θ
26 Decision Rules Measurements Likelihood function Prior Distribution Posterior Distrribution Gain Function Result Slide 26/31 θ
27 Some Desirable Properties of Estimators I: Unbiased: Mean value of the error should be zero E Φ ) - Φ = 0 E Φ ) = E θ Consistent: Error estimator should decrease asymptotically as number of measurements increase. (Mean Square Error (MSE)) MSE ) 2 = E Φ - Φ 0 for large What happens to MSE when estimator is biased? 2 2 MSE = E Φ ) - Φ - b + E b variance Slide 27/31 bias
28 Some Desirable Properties of Estimators II: Efficient: Co-variance matrix of error should decrease asymptotically to its minimal value for large T 1 ( Φi - i)( Φk - k) J ik ) ) C % = E Φ Φ θ Slide 28/31
29 Example: Properties Of Estimators Mean and Variance Mean: 1 1 E ˆ μ = E ( j) ϕ = μ = μ j= 1 The sample mean is an unbiased estimator of the true mean Variance: σ E ( ˆ μ μ) = E ϕ( j) μ = σ = ( ) 2 2 j= 1 The variance is a consistent estimator because It approaches zero for large number of measurements. Slide 29/31
30 Properties Of MLE is consistent: the MLE recovers asymptotically the true parameter values that generated the data for inf; Is efficient: The MLE achieves asymptotically the minimum error (= max. information) Slide 30/31
31 Summary LSE is a descriptive method to accurately fit data to a model. MLE is a method to seek the probability distribution that makes the observed data most likely. MAP is a method to seek the most probably parameter value given prior information about the parameters and the observed data. If the influence of prior information decreases, i.e. many measurements, MAP approaches MLE Slide 31/31
32 Some Priors in Imaging Smoothness of the brain Anatomical boundaries Intensity distributions Anatomical shapes Physical models Point spread function Bandwidth limits Etc. Slide 32/31
33 Estimation Theory: Motivation Example I Gray/White Matter Segmentation Hypothetical Histogram Intensity What works better than flipping a coin? Design likelihood functions based on anatomy co-occurance of signal intensities others Determine prior distribution population based atlas of regional intensities model based distributions of intensities others Slide 33/31
34 Estimation Theory: Motivation Example II Goal: Capture dynamic signal on a static background Poor signal to noise Improvements to identify the dynamic signal: D. Feinberg Advanced MRI Technologies, Sebastopol, CA Slide 34/31 Design likelihood functions based on auto-correlations anatomical information Determine prior distributions from serial measurements multiple subjects anatomy
35 Estimation Theory: Motivation Example III Diffusion Spectrum Imaging Human Cingulum Bundle Goal: Capture directions of fiber bundles Improvements to identify tracts: Design likelihood functions based on similarity measures of adjacent voxels fiber anatomy Determine prior distributions from anatomy fiber skeletons from a population others Dr. Van Wedeen, MGH Slide 35/31
36 MAP Estimation in Image Reconstructions with Edge-Preserving Priors Slide 36/31 Dr. Ashish Raj, Cornell U
37 MAP Estimation In Image Reconstruction Human brain MRI. (a) The original LR data. (b) Zero-padding interpolation. (c) SR with box-psf. (d) SR with Gaussian-PSF. From: A. Greenspan in The Computer Journal Advance Access published February 19, 2008 Slide 37/31
38 Improved ASL Perfusion Results zdft = zero-filled DFT By Dr. John Kornak, UCSF
39 Bayesian Automated Image Segmentation Bruce Fischl, MGH Slide 39/31
40 Segmentation Using MLE A: Raw MRI B: SPM2 C: EMS D: HBSA from Habib Zaidi, et al, euroimage 32 (2006) Slide 40/31
41 Population Atlases As Priors Slide 41/31 Dr. Sarang Joshi, U Utah, Salt Lake City
42 Population Shape Regressions Based Age- Selective Priors Age = Slide 42/31 Dr. Sarang Joshi, U Utah, Salt Lake City
43 Imaging Software Using MLE And MAP Packages Applications Languages VoxBo fmri C/C++/IDL MEDx smri, fmri C/C++/Tcl/Tk SPM fmri, smri matlab/c ibrain IDL FSL fmri, smri, DTI C/C++ fmristat fmri matlab BrainVoyager smri C/C++ BrainTools C/C++ AFI fmri, DTI C/C++ Freesurfer smri C/C++ ipy Python Slide 43/31
44 Literature Mathematical H. Sorenson. Parameter Estimation Principles and Problems. Marcel Dekker (pub)1980. Signal Processing S. Kay. Fundamentals of Signal Processing Estimation Theory. Prentice Hall L. Scharf. Statistical Signal Processing: Detection, Estimation, and Time Series Analysis. Addison-Wesley Statistics: A. Hyvarinen. Independent Component Analysis. John Wileys & Sons ew Directions in Statistical Signal Processing. From Systems to Brain. Ed. S. Haykin. MIT Press Slide 44/31
Parametric Models. Dr. Shuang LIANG. School of Software Engineering TongJi University Fall, 2012
Parametric Models Dr. Shuang LIANG School of Software Engineering TongJi University Fall, 2012 Today s Topics Maximum Likelihood Estimation Bayesian Density Estimation Today s Topics Maximum Likelihood
More informationStatistics: Learning models from data
DS-GA 1002 Lecture notes 5 October 19, 2015 Statistics: Learning models from data Learning models from data that are assumed to be generated probabilistically from a certain unknown distribution is a crucial
More informationMS&E 226: Small Data. Lecture 11: Maximum likelihood (v2) Ramesh Johari
MS&E 226: Small Data Lecture 11: Maximum likelihood (v2) Ramesh Johari ramesh.johari@stanford.edu 1 / 18 The likelihood function 2 / 18 Estimating the parameter This lecture develops the methodology behind
More informationEIE6207: Maximum-Likelihood and Bayesian Estimation
EIE6207: Maximum-Likelihood and Bayesian Estimation Man-Wai MAK Dept. of Electronic and Information Engineering, The Hong Kong Polytechnic University enmwmak@polyu.edu.hk http://www.eie.polyu.edu.hk/ mwmak
More informationStatistical and Learning Techniques in Computer Vision Lecture 2: Maximum Likelihood and Bayesian Estimation Jens Rittscher and Chuck Stewart
Statistical and Learning Techniques in Computer Vision Lecture 2: Maximum Likelihood and Bayesian Estimation Jens Rittscher and Chuck Stewart 1 Motivation and Problem In Lecture 1 we briefly saw how histograms
More informationBayesian Inference and MCMC
Bayesian Inference and MCMC Aryan Arbabi Partly based on MCMC slides from CSC412 Fall 2018 1 / 18 Bayesian Inference - Motivation Consider we have a data set D = {x 1,..., x n }. E.g each x i can be the
More informationParametric Techniques Lecture 3
Parametric Techniques Lecture 3 Jason Corso SUNY at Buffalo 22 January 2009 J. Corso (SUNY at Buffalo) Parametric Techniques Lecture 3 22 January 2009 1 / 39 Introduction In Lecture 2, we learned how to
More informationModule 2. Random Processes. Version 2, ECE IIT, Kharagpur
Module Random Processes Version, ECE IIT, Kharagpur Lesson 9 Introduction to Statistical Signal Processing Version, ECE IIT, Kharagpur After reading this lesson, you will learn about Hypotheses testing
More informationLecture : Probabilistic Machine Learning
Lecture : Probabilistic Machine Learning Riashat Islam Reasoning and Learning Lab McGill University September 11, 2018 ML : Many Methods with Many Links Modelling Views of Machine Learning Machine Learning
More informationCSC321 Lecture 18: Learning Probabilistic Models
CSC321 Lecture 18: Learning Probabilistic Models Roger Grosse Roger Grosse CSC321 Lecture 18: Learning Probabilistic Models 1 / 25 Overview So far in this course: mainly supervised learning Language modeling
More informationA Brief Review of Probability, Bayesian Statistics, and Information Theory
A Brief Review of Probability, Bayesian Statistics, and Information Theory Brendan Frey Electrical and Computer Engineering University of Toronto frey@psi.toronto.edu http://www.psi.toronto.edu A system
More informationParametric Techniques
Parametric Techniques Jason J. Corso SUNY at Buffalo J. Corso (SUNY at Buffalo) Parametric Techniques 1 / 39 Introduction When covering Bayesian Decision Theory, we assumed the full probabilistic structure
More informationQuantitative Neuro-Anatomic and Functional Image Assessment Recent progress on image registration and its applications
Quantitative Neuro-Anatomic and Functional Image Assessment Recent progress on image registration and its applications Guido Gerig Sarang Joshi Tom Fletcher Applications of image registration in neuroimaging
More informationMachine Learning. Lecture 4: Regularization and Bayesian Statistics. Feng Li. https://funglee.github.io
Machine Learning Lecture 4: Regularization and Bayesian Statistics Feng Li fli@sdu.edu.cn https://funglee.github.io School of Computer Science and Technology Shandong University Fall 207 Overfitting Problem
More informationDensity Estimation: ML, MAP, Bayesian estimation
Density Estimation: ML, MAP, Bayesian estimation CE-725: Statistical Pattern Recognition Sharif University of Technology Spring 2013 Soleymani Outline Introduction Maximum-Likelihood Estimation Maximum
More informationIntroduction to Machine Learning. Maximum Likelihood and Bayesian Inference. Lecturers: Eran Halperin, Lior Wolf
1 Introduction to Machine Learning Maximum Likelihood and Bayesian Inference Lecturers: Eran Halperin, Lior Wolf 2014-15 We know that X ~ B(n,p), but we do not know p. We get a random sample from X, a
More informationDS-GA 1002 Lecture notes 11 Fall Bayesian statistics
DS-GA 100 Lecture notes 11 Fall 016 Bayesian statistics In the frequentist paradigm we model the data as realizations from a distribution that depends on deterministic parameters. In contrast, in Bayesian
More informationMachine Learning Basics: Maximum Likelihood Estimation
Machine Learning Basics: Maximum Likelihood Estimation Sargur N. srihari@cedar.buffalo.edu This is part of lecture slides on Deep Learning: http://www.cedar.buffalo.edu/~srihari/cse676 1 Topics 1. Learning
More informationDS-GA 1003: Machine Learning and Computational Statistics Homework 7: Bayesian Modeling
DS-GA 1003: Machine Learning and Computational Statistics Homework 7: Bayesian Modeling Due: Tuesday, May 10, 2016, at 6pm (Submit via NYU Classes) Instructions: Your answers to the questions below, including
More informationIntroduction: MLE, MAP, Bayesian reasoning (28/8/13)
STA561: Probabilistic machine learning Introduction: MLE, MAP, Bayesian reasoning (28/8/13) Lecturer: Barbara Engelhardt Scribes: K. Ulrich, J. Subramanian, N. Raval, J. O Hollaren 1 Classifiers In this
More informationParameter Estimation. Industrial AI Lab.
Parameter Estimation Industrial AI Lab. Generative Model X Y w y = ω T x + ε ε~n(0, σ 2 ) σ 2 2 Maximum Likelihood Estimation (MLE) Estimate parameters θ ω, σ 2 given a generative model Given observed
More information2 Statistical Estimation: Basic Concepts
Technion Israel Institute of Technology, Department of Electrical Engineering Estimation and Identification in Dynamical Systems (048825) Lecture Notes, Fall 2009, Prof. N. Shimkin 2 Statistical Estimation:
More informationNaïve Bayes classification
Naïve Bayes classification 1 Probability theory Random variable: a variable whose possible values are numerical outcomes of a random phenomenon. Examples: A person s height, the outcome of a coin toss
More informationCOMP90051 Statistical Machine Learning
COMP90051 Statistical Machine Learning Semester 2, 2017 Lecturer: Trevor Cohn 2. Statistical Schools Adapted from slides by Ben Rubinstein Statistical Schools of Thought Remainder of lecture is to provide
More informationDetection theory. H 0 : x[n] = w[n]
Detection Theory Detection theory A the last topic of the course, we will briefly consider detection theory. The methods are based on estimation theory and attempt to answer questions such as Is a signal
More informationBayesian Methods: Naïve Bayes
Bayesian Methods: aïve Bayes icholas Ruozzi University of Texas at Dallas based on the slides of Vibhav Gogate Last Time Parameter learning Learning the parameter of a simple coin flipping model Prior
More informationChapter 3: Maximum-Likelihood & Bayesian Parameter Estimation (part 1)
HW 1 due today Parameter Estimation Biometrics CSE 190 Lecture 7 Today s lecture was on the blackboard. These slides are an alternative presentation of the material. CSE190, Winter10 CSE190, Winter10 Chapter
More informationSTATS 200: Introduction to Statistical Inference. Lecture 29: Course review
STATS 200: Introduction to Statistical Inference Lecture 29: Course review Course review We started in Lecture 1 with a fundamental assumption: Data is a realization of a random process. The goal throughout
More informationEEL 851: Biometrics. An Overview of Statistical Pattern Recognition EEL 851 1
EEL 851: Biometrics An Overview of Statistical Pattern Recognition EEL 851 1 Outline Introduction Pattern Feature Noise Example Problem Analysis Segmentation Feature Extraction Classification Design Cycle
More informationAccelerated MRI Image Reconstruction
IMAGING DATA EVALUATION AND ANALYTICS LAB (IDEAL) CS5540: Computational Techniques for Analyzing Clinical Data Lecture 15: Accelerated MRI Image Reconstruction Ashish Raj, PhD Image Data Evaluation and
More informationTutorial on Probabilistic Programming with PyMC3
185.A83 Machine Learning for Health Informatics 2017S, VU, 2.0 h, 3.0 ECTS Tutorial 02-04.04.2017 Tutorial on Probabilistic Programming with PyMC3 florian.endel@tuwien.ac.at http://hci-kdd.org/machine-learning-for-health-informatics-course
More informationDensity Estimation. Seungjin Choi
Density Estimation Seungjin Choi Department of Computer Science and Engineering Pohang University of Science and Technology 77 Cheongam-ro, Nam-gu, Pohang 37673, Korea seungjin@postech.ac.kr http://mlg.postech.ac.kr/
More informationProbability and Estimation. Alan Moses
Probability and Estimation Alan Moses Random variables and probability A random variable is like a variable in algebra (e.g., y=e x ), but where at least part of the variability is taken to be stochastic.
More informationIntroduction to Bayesian Inference
University of Pennsylvania EABCN Training School May 10, 2016 Bayesian Inference Ingredients of Bayesian Analysis: Likelihood function p(y φ) Prior density p(φ) Marginal data density p(y ) = p(y φ)p(φ)dφ
More informationBayesian Inference. Introduction
Bayesian Inference Introduction The frequentist approach to inference holds that probabilities are intrinsicially tied (unsurprisingly) to frequencies. This interpretation is actually quite natural. What,
More informationDetection theory 101 ELEC-E5410 Signal Processing for Communications
Detection theory 101 ELEC-E5410 Signal Processing for Communications Binary hypothesis testing Null hypothesis H 0 : e.g. noise only Alternative hypothesis H 1 : signal + noise p(x;h 0 ) γ p(x;h 1 ) Trade-off
More informationNaïve Bayes classification. p ij 11/15/16. Probability theory. Probability theory. Probability theory. X P (X = x i )=1 i. Marginal Probability
Probability theory Naïve Bayes classification Random variable: a variable whose possible values are numerical outcomes of a random phenomenon. s: A person s height, the outcome of a coin toss Distinguish
More informationOutline Lecture 2 2(32)
Outline Lecture (3), Lecture Linear Regression and Classification it is our firm belief that an understanding of linear models is essential for understanding nonlinear ones Thomas Schön Division of Automatic
More informationLecture 11: Probability Distributions and Parameter Estimation
Intelligent Data Analysis and Probabilistic Inference Lecture 11: Probability Distributions and Parameter Estimation Recommended reading: Bishop: Chapters 1.2, 2.1 2.3.4, Appendix B Duncan Gillies and
More informationLearning From Crowds. Presented by: Bei Peng 03/24/15
Learning From Crowds Presented by: Bei Peng 03/24/15 1 Supervised Learning Given labeled training data, learn to generalize well on unseen data Binary classification ( ) Multi-class classification ( y
More informationMachine Learning
Machine Learning 10-701 Tom M. Mitchell Machine Learning Department Carnegie Mellon University February 1, 2011 Today: Generative discriminative classifiers Linear regression Decomposition of error into
More informationLecture 18: Bayesian Inference
Lecture 18: Bayesian Inference Hyang-Won Lee Dept. of Internet & Multimedia Eng. Konkuk University Lecture 18 Probability and Statistics, Spring 2014 1 / 10 Bayesian Statistical Inference Statiscal inference
More informationBayesian Learning (II)
Universität Potsdam Institut für Informatik Lehrstuhl Maschinelles Lernen Bayesian Learning (II) Niels Landwehr Overview Probabilities, expected values, variance Basic concepts of Bayesian learning MAP
More informationIntroduction to Probabilistic Machine Learning
Introduction to Probabilistic Machine Learning Piyush Rai Dept. of CSE, IIT Kanpur (Mini-course 1) Nov 03, 2015 Piyush Rai (IIT Kanpur) Introduction to Probabilistic Machine Learning 1 Machine Learning
More informationCS-E3210 Machine Learning: Basic Principles
CS-E3210 Machine Learning: Basic Principles Lecture 4: Regression II slides by Markus Heinonen Department of Computer Science Aalto University, School of Science Autumn (Period I) 2017 1 / 61 Today s introduction
More informationMachine Learning CSE546 Carlos Guestrin University of Washington. September 30, 2013
Bayesian Methods Machine Learning CSE546 Carlos Guestrin University of Washington September 30, 2013 1 What about prior n Billionaire says: Wait, I know that the thumbtack is close to 50-50. What can you
More informationECE521 week 3: 23/26 January 2017
ECE521 week 3: 23/26 January 2017 Outline Probabilistic interpretation of linear regression - Maximum likelihood estimation (MLE) - Maximum a posteriori (MAP) estimation Bias-variance trade-off Linear
More informationAccouncements. You should turn in a PDF and a python file(s) Figure for problem 9 should be in the PDF
Accouncements You should turn in a PDF and a python file(s) Figure for problem 9 should be in the PDF Please do not zip these files and submit (unless there are >5 files) 1 Bayesian Methods Machine Learning
More informationECE531 Lecture 10b: Maximum Likelihood Estimation
ECE531 Lecture 10b: Maximum Likelihood Estimation D. Richard Brown III Worcester Polytechnic Institute 05-Apr-2011 Worcester Polytechnic Institute D. Richard Brown III 05-Apr-2011 1 / 23 Introduction So
More informationNaïve Bayes Introduction to Machine Learning. Matt Gormley Lecture 18 Oct. 31, 2018
10-601 Introduction to Machine Learning Machine Learning Department School of Computer Science Carnegie Mellon University Naïve Bayes Matt Gormley Lecture 18 Oct. 31, 2018 1 Reminders Homework 6: PAC Learning
More informationIntroduction to Machine Learning. Maximum Likelihood and Bayesian Inference. Lecturers: Eran Halperin, Yishay Mansour, Lior Wolf
1 Introduction to Machine Learning Maximum Likelihood and Bayesian Inference Lecturers: Eran Halperin, Yishay Mansour, Lior Wolf 2013-14 We know that X ~ B(n,p), but we do not know p. We get a random sample
More informationMachine Learning. Gaussian Mixture Models. Zhiyao Duan & Bryan Pardo, Machine Learning: EECS 349 Fall
Machine Learning Gaussian Mixture Models Zhiyao Duan & Bryan Pardo, Machine Learning: EECS 349 Fall 2012 1 The Generative Model POV We think of the data as being generated from some process. We assume
More informationIntroduction to Gaussian Process
Introduction to Gaussian Process CS 778 Chris Tensmeyer CS 478 INTRODUCTION 1 What Topic? Machine Learning Regression Bayesian ML Bayesian Regression Bayesian Non-parametric Gaussian Process (GP) GP Regression
More informationComposite Hypotheses and Generalized Likelihood Ratio Tests
Composite Hypotheses and Generalized Likelihood Ratio Tests Rebecca Willett, 06 In many real world problems, it is difficult to precisely specify probability distributions. Our models for data may involve
More informationPoint Estimation. Vibhav Gogate The University of Texas at Dallas
Point Estimation Vibhav Gogate The University of Texas at Dallas Some slides courtesy of Carlos Guestrin, Chris Bishop, Dan Weld and Luke Zettlemoyer. Basics: Expectation and Variance Binary Variables
More informationBasics of Diffusion Tensor Imaging and DtiStudio
Basics of Diffusion Tensor Imaging and DtiStudio DTI Basics 1 DTI reveals White matter anatomy Gray matter White matter DTI uses water diffusion as a probe for white matter anatomy Isotropic diffusion
More informationProbabilistic modeling. The slides are closely adapted from Subhransu Maji s slides
Probabilistic modeling The slides are closely adapted from Subhransu Maji s slides Overview So far the models and algorithms you have learned about are relatively disconnected Probabilistic modeling framework
More informationMachine Learning
Machine Learning 10-601 Tom M. Mitchell Machine Learning Department Carnegie Mellon University February 4, 2015 Today: Generative discriminative classifiers Linear regression Decomposition of error into
More informationHierarchical Models & Bayesian Model Selection
Hierarchical Models & Bayesian Model Selection Geoffrey Roeder Departments of Computer Science and Statistics University of British Columbia Jan. 20, 2016 Contact information Please report any typos or
More informationCLASS NOTES Models, Algorithms and Data: Introduction to computing 2018
CLASS NOTES Models, Algorithms and Data: Introduction to computing 208 Petros Koumoutsakos, Jens Honore Walther (Last update: June, 208) IMPORTANT DISCLAIMERS. REFERENCES: Much of the material (ideas,
More informationModule 1 - Signal estimation
, Arraial do Cabo, 2009 Module 1 - Signal estimation Sérgio M. Jesus (sjesus@ualg.pt) Universidade do Algarve, PT-8005-139 Faro, Portugal www.siplab.fct.ualg.pt February 2009 Outline of Module 1 Parameter
More informationFrom Diffusion Data to Bundle Analysis
From Diffusion Data to Bundle Analysis Gabriel Girard gabriel.girard@epfl.ch Computational Brain Connectivity Mapping Juan-les-Pins, France 20 November 2017 Gabriel Girard gabriel.girard@epfl.ch CoBCoM2017
More informationMaximum Likelihood, Logistic Regression, and Stochastic Gradient Training
Maximum Likelihood, Logistic Regression, and Stochastic Gradient Training Charles Elkan elkan@cs.ucsd.edu January 17, 2013 1 Principle of maximum likelihood Consider a family of probability distributions
More information+ + ( + ) = Linear recurrent networks. Simpler, much more amenable to analytic treatment E.g. by choosing
Linear recurrent networks Simpler, much more amenable to analytic treatment E.g. by choosing + ( + ) = Firing rates can be negative Approximates dynamics around fixed point Approximation often reasonable
More informationECE 521. Lecture 11 (not on midterm material) 13 February K-means clustering, Dimensionality reduction
ECE 521 Lecture 11 (not on midterm material) 13 February 2017 K-means clustering, Dimensionality reduction With thanks to Ruslan Salakhutdinov for an earlier version of the slides Overview K-means clustering
More informationSome slides from Carlos Guestrin, Luke Zettlemoyer & K Gajos 2
Logistics CSE 446: Point Estimation Winter 2012 PS2 out shortly Dan Weld Some slides from Carlos Guestrin, Luke Zettlemoyer & K Gajos 2 Last Time Random variables, distributions Marginal, joint & conditional
More informationMINIMUM EXPECTED RISK PROBABILITY ESTIMATES FOR NONPARAMETRIC NEIGHBORHOOD CLASSIFIERS. Maya Gupta, Luca Cazzanti, and Santosh Srivastava
MINIMUM EXPECTED RISK PROBABILITY ESTIMATES FOR NONPARAMETRIC NEIGHBORHOOD CLASSIFIERS Maya Gupta, Luca Cazzanti, and Santosh Srivastava University of Washington Dept. of Electrical Engineering Seattle,
More informationCHAPTER 2 Estimating Probabilities
CHAPTER 2 Estimating Probabilities Machine Learning Copyright c 2017. Tom M. Mitchell. All rights reserved. *DRAFT OF September 16, 2017* *PLEASE DO NOT DISTRIBUTE WITHOUT AUTHOR S PERMISSION* This is
More informationLecture 2: From Linear Regression to Kalman Filter and Beyond
Lecture 2: From Linear Regression to Kalman Filter and Beyond January 18, 2017 Contents 1 Batch and Recursive Estimation 2 Towards Bayesian Filtering 3 Kalman Filter and Bayesian Filtering and Smoothing
More informationLecture 2: Statistical Decision Theory (Part I)
Lecture 2: Statistical Decision Theory (Part I) Hao Helen Zhang Hao Helen Zhang Lecture 2: Statistical Decision Theory (Part I) 1 / 35 Outline of This Note Part I: Statistics Decision Theory (from Statistical
More informationOverview. Probabilistic Interpretation of Linear Regression Maximum Likelihood Estimation Bayesian Estimation MAP Estimation
Overview Probabilistic Interpretation of Linear Regression Maximum Likelihood Estimation Bayesian Estimation MAP Estimation Probabilistic Interpretation: Linear Regression Assume output y is generated
More informationEIE6207: Estimation Theory
EIE6207: Estimation Theory Man-Wai MAK Dept. of Electronic and Information Engineering, The Hong Kong Polytechnic University enmwmak@polyu.edu.hk http://www.eie.polyu.edu.hk/ mwmak References: Steven M.
More informationCourse content (will be adapted to the background knowledge of the class):
Biomedical Signal Processing and Signal Modeling Lucas C Parra, parra@ccny.cuny.edu Departamento the Fisica, UBA Synopsis This course introduces two fundamental concepts of signal processing: linear systems
More informationLanguage Modelling: Smoothing and Model Complexity. COMP-599 Sept 14, 2016
Language Modelling: Smoothing and Model Complexity COMP-599 Sept 14, 2016 Announcements A1 has been released Due on Wednesday, September 28th Start code for Question 4: Includes some of the package import
More informationLecture 8: Information Theory and Statistics
Lecture 8: Information Theory and Statistics Part II: Hypothesis Testing and I-Hsiang Wang Department of Electrical Engineering National Taiwan University ihwang@ntu.edu.tw December 23, 2015 1 / 50 I-Hsiang
More informationStatistical Pattern Recognition
Statistical Pattern Recognition Expectation Maximization (EM) and Mixture Models Hamid R. Rabiee Jafar Muhammadi, Mohammad J. Hosseini Spring 2014 http://ce.sharif.edu/courses/92-93/2/ce725-2 Agenda Expectation-maximization
More informationCOMS 4721: Machine Learning for Data Science Lecture 1, 1/17/2017
COMS 4721: Machine Learning for Data Science Lecture 1, 1/17/2017 Prof. John Paisley Department of Electrical Engineering & Data Science Institute Columbia University OVERVIEW This class will cover model-based
More informationComputational Perception. Bayesian Inference
Computational Perception 15-485/785 January 24, 2008 Bayesian Inference The process of probabilistic inference 1. define model of problem 2. derive posterior distributions and estimators 3. estimate parameters
More information1/24/2008. Review of Statistical Inference. C.1 A Sample of Data. C.2 An Econometric Model. C.4 Estimating the Population Variance and Other Moments
/4/008 Review of Statistical Inference Prepared by Vera Tabakova, East Carolina University C. A Sample of Data C. An Econometric Model C.3 Estimating the Mean of a Population C.4 Estimating the Population
More informationHomework 5: Conditional Probability Models
Homework 5: Conditional Probability Models Instructions: Your answers to the questions below, including plots and mathematical work, should be submitted as a single PDF file. It s preferred that you write
More informationLearning Bayesian network : Given structure and completely observed data
Learning Bayesian network : Given structure and completely observed data Probabilistic Graphical Models Sharif University of Technology Spring 2017 Soleymani Learning problem Target: true distribution
More informationTheory of Maximum Likelihood Estimation. Konstantin Kashin
Gov 2001 Section 5: Theory of Maximum Likelihood Estimation Konstantin Kashin February 28, 2013 Outline Introduction Likelihood Examples of MLE Variance of MLE Asymptotic Properties What is Statistical
More informationCOS513: FOUNDATIONS OF PROBABILISTIC MODELS LECTURE 9: LINEAR REGRESSION
COS513: FOUNDATIONS OF PROBABILISTIC MODELS LECTURE 9: LINEAR REGRESSION SEAN GERRISH AND CHONG WANG 1. WAYS OF ORGANIZING MODELS In probabilistic modeling, there are several ways of organizing models:
More informationAn Introduction to Statistical and Probabilistic Linear Models
An Introduction to Statistical and Probabilistic Linear Models Maximilian Mozes Proseminar Data Mining Fakultät für Informatik Technische Universität München June 07, 2017 Introduction In statistical learning
More informationMachine Learning 4771
Machine Learning 4771 Instructor: Tony Jebara Topic 11 Maximum Likelihood as Bayesian Inference Maximum A Posteriori Bayesian Gaussian Estimation Why Maximum Likelihood? So far, assumed max (log) likelihood
More informationMS&E 226: Small Data
MS&E 226: Small Data Lecture 12: Frequentist properties of estimators (v4) Ramesh Johari ramesh.johari@stanford.edu 1 / 39 Frequentist inference 2 / 39 Thinking like a frequentist Suppose that for some
More informationECE521 Lecture7. Logistic Regression
ECE521 Lecture7 Logistic Regression Outline Review of decision theory Logistic regression A single neuron Multi-class classification 2 Outline Decision theory is conceptually easy and computationally hard
More informationVariations. ECE 6540, Lecture 10 Maximum Likelihood Estimation
Variations ECE 6540, Lecture 10 Last Time BLUE (Best Linear Unbiased Estimator) Formulation Advantages Disadvantages 2 The BLUE A simplification Assume the estimator is a linear system For a single parameter
More informationLecture 5: GPs and Streaming regression
Lecture 5: GPs and Streaming regression Gaussian Processes Information gain Confidence intervals COMP-652 and ECSE-608, Lecture 5 - September 19, 2017 1 Recall: Non-parametric regression Input space X
More informationECE521 Tutorial 11. Topic Review. ECE521 Winter Credits to Alireza Makhzani, Alex Schwing, Rich Zemel and TAs for slides. ECE521 Tutorial 11 / 4
ECE52 Tutorial Topic Review ECE52 Winter 206 Credits to Alireza Makhzani, Alex Schwing, Rich Zemel and TAs for slides ECE52 Tutorial ECE52 Winter 206 Credits to Alireza / 4 Outline K-means, PCA 2 Bayesian
More informationIEOR E4570: Machine Learning for OR&FE Spring 2015 c 2015 by Martin Haugh. The EM Algorithm
IEOR E4570: Machine Learning for OR&FE Spring 205 c 205 by Martin Haugh The EM Algorithm The EM algorithm is used for obtaining maximum likelihood estimates of parameters when some of the data is missing.
More informationCOS513 LECTURE 8 STATISTICAL CONCEPTS
COS513 LECTURE 8 STATISTICAL CONCEPTS NIKOLAI SLAVOV AND ANKUR PARIKH 1. MAKING MEANINGFUL STATEMENTS FROM JOINT PROBABILITY DISTRIBUTIONS. A graphical model (GM) represents a family of probability distributions
More informationIntroduction to Bayesian Inference
Introduction to Bayesian Inference p. 1/2 Introduction to Bayesian Inference September 15th, 2010 Reading: Hoff Chapter 1-2 Introduction to Bayesian Inference p. 2/2 Probability: Measurement of Uncertainty
More informationMachine Learning CSE546 Carlos Guestrin University of Washington. September 30, What about continuous variables?
Linear Regression Machine Learning CSE546 Carlos Guestrin University of Washington September 30, 2014 1 What about continuous variables? n Billionaire says: If I am measuring a continuous variable, what
More informationBayesian inference J. Daunizeau
Bayesian inference J. Daunizeau Brain and Spine Institute, Paris, France Wellcome Trust Centre for Neuroimaging, London, UK Overview of the talk 1 Probabilistic modelling and representation of uncertainty
More informationStatistics - Lecture One. Outline. Charlotte Wickham 1. Basic ideas about estimation
Statistics - Lecture One Charlotte Wickham wickham@stat.berkeley.edu http://www.stat.berkeley.edu/~wickham/ Outline 1. Basic ideas about estimation 2. Method of Moments 3. Maximum Likelihood 4. Confidence
More informationEstimation Theory. as Θ = (Θ 1,Θ 2,...,Θ m ) T. An estimator
Estimation Theory Estimation theory deals with finding numerical values of interesting parameters from given set of data. We start with formulating a family of models that could describe how the data were
More informationDouble Gamma Principal Components Analysis
Applied Mathematical Sciences, Vol. 12, 2018, no. 11, 523-533 HIKARI Ltd, www.m-hikari.com https://doi.org/10.12988/ams.2018.8455 Double Gamma Principal Components Analysis Ameerah O. Bahashwan, Zakiah
More informationLecture 3. G. Cowan. Lecture 3 page 1. Lectures on Statistical Data Analysis
Lecture 3 1 Probability (90 min.) Definition, Bayes theorem, probability densities and their properties, catalogue of pdfs, Monte Carlo 2 Statistical tests (90 min.) general concepts, test statistics,
More informationIntroduction to Systems Analysis and Decision Making Prepared by: Jakub Tomczak
Introduction to Systems Analysis and Decision Making Prepared by: Jakub Tomczak 1 Introduction. Random variables During the course we are interested in reasoning about considered phenomenon. In other words,
More information