Statistical Estimation: Least Squares, Maximum Likelihood and Maximum A Posteriori Estimators

Size: px
Start display at page:

Download "Statistical Estimation: Least Squares, Maximum Likelihood and Maximum A Posteriori Estimators"

Transcription

1 CS5540: Computatonal Technques for Analyzng Clncal Data Lecture 7: Statstcal Estmaton: Least Squares, Maxmum Lkelhood and Maxmum A Posteror Estmators Ashsh Raj, PhD Image Data Evaluaton and Analytcs Laboratory (IDEAL) Department of Radology Well Cornell Medcal College New York

2 Outlne Part I: Recap of Wavelet Transforms Part II : Least Squares Estmaton Part III: Maxmum Lkelhood Estmaton Part IV: Maxmum A Posteror Estmaton : Next week Note: you wll not be tested on specfc examples shown here, only on general prncples IDEA Lab, Radology, Cornell 2

3 Bass functons n WT Haar Mexcan Hat Daubeches (orthonormal) IDEA Lab, Radology, Cornell 3

4 WT n mages Images are pecewse smooth or pecewse constant Statonarty s even rarer than n 1D sgnals FT even less useful (nnd WT more attractve) 2D wavelet transforms are smple extensons of 1D WT, generally performng 1D WT along rows, then columns etc Sometmes we use 2D wavelets drectly, e.g. orthonormal Daubeches 2D wavelet IDEA Lab, Radology, Cornell 4

5 WT on mages 2D generalzaton of scaletme decomposton scale Scale 0 Scale 1 H tme Scale 2 Successve applcaton of dot product wth wavelet of ncreasng wdth. Forms a natural pyramd structure. At each scale: H = dot product of mage rows wth wavelet V = dot product of mage columns wth wavelet H-V = dot product of mage rows then columns wth wavelet IDEA Lab, Radology, Cornell V H-V 5

6 Wavelet Applcatons Many, many applcatons! Audo, mage and vdeo compresson New JPEG standard ncludes wavelet compresson FBI s fngerprnts database saved as waveletcompressed Sgnal denosng, nterpolaton, mage zoomng, texture analyss, tme-scale feature extracton In our context, WT wll be used prmarly as a feature extracton tool Remember, WT s just a change of bass, n order to extract useful nformaton whch mght otherwse not be easly seen IDEA Lab, Radology, Cornell 6

7 WT n MATLAB MATLAB has an extensve wavelet toolbox Type help wavelet n MATLAB command wndow Look at ther wavelet demo Play wth Haar, Mexcan hat and Daubeches wavelets IDEA Lab, Radology, Cornell 7

8 Project Ideas Idea 1: use WT to extract features from ECG data use these features for classfcaton Idea 2: use 2D WT to extract spato-temporal features from 3D+tme MRI data to detect tumors / classfy bengn vs malgnant tumors Idea 3: use 2D WT to denose a gven mage IDEA Lab, Radology, Cornell 8

9 Idea 3: Voxel labelng from contrast-enhanced MRI Can segment accordng to tme profle of 3D+tme contrast enhanced MR data of lver / mammography Typcal plot of tme-resolved MR sgnal of varous tssue classes Temporal models used to extract features Instead of such a smple temporal model, wavelet decomposton could provde spato-temporal features that you can use for clusterng IDEA Lab, Radology, Cornell

10 Lver tumour quantfcaton from DCE-MRI IDEA Lab, Radology, Cornell

11 Further Readng on Wavelets IDEA Lab, Radology, Cornell 11

12 Part II : Least Squares Estmaton and Examples

13 A smple Least Squares problem Lne fttng Goal: To fnd the best-ft lne representng a bunch of ponts Here: y are observatons at locaton x, Intercept and slope of lne are the unknown model parameters to be estmated Whch model parameters best ft the observed ponts? E( y nt Best ( y, m) nt =, m) ( y h( x where, m) Ths can be wrtten n matrx notaton, as θ LS = arg mn y Hθ 2 )) 2 = arg mn E( y, nt h( x) = y nt + mx What are H and θ? IDEA Lab, Radology, Cornell 13

14 Least Squares Estmator Gven lnear process y = H θ + n Least Squares estmator: θ LS = argmn y Hθ 2 Natural estmator want soluton to match observaton Does not use any nformaton about n There s a smple soluton (a.k.a. pseudo-nverse): θ LS = (H T H) -1 H T y In MATLAB, type pnv(y) IDEA Lab, Radology, Cornell 14

15 Example - estmatng T 2 decay constant n repeated spn echo MR data IDEA Lab, Radology, Cornell 15

16 Example estmatng T 2 n repeated spn echo data s(t) = s 0 e -t/t 2 Need only 2 data ponts to estmate T 2 : T 2est = [T E2 T E1 ] / ln[s(t E1 )/s(t E2 ) ] However, not good due to nose, tmng ssues In practce we have many data samples from varous echoes IDEA Lab, Radology, Cornell 16

17 Example estmatng T 2 y ln(s(t 1 )) ln(s(t 2 )) M ln(s(t n )) = 1 -t 1 1 -t 2 M 1 -t n a r H θ Least Squares estmate: θ LS = (H T H) -1 H T y T 2 = 1/r LS IDEA Lab, Radology, Cornell 17

18 Estmaton example - Denosng Suppose we have a nosy MR mage y, and wsh to obtan the noseless mage x, where y = x + n Can we use LSE to fnd x? Try: H = I, θ = x n the lnear model LS estmator smply gves x = y! we need a more powerful model Suppose the mage x can be approxmated by a polynomal,.e. a mxture of 1 st p powers of r: x = Σ =0 p a r IDEA Lab, Radology, Cornell 18

19 y Example denosng y 1 y 2 M y n = 1 r 1 1 L r 1 p 1 r 2 1 L r 2 p M 1 r n 1 L r n p a 0 a 1 M a p H + n 1 n 2 M n n Least Squares estmate: θ LS = (H T H) -1 H T y θ x = Σ =0 p a r IDEA Lab, Radology, Cornell 19

20 Part III : Maxmum Lkelhood Estmaton and Examples

21 Estmaton Theory Consder a lnear process y = H θ + n y = observed data θ = set of model parameters n = addtve nose Then Estmaton s the problem of fndng the statstcally optmal θ, gven y, H and knowledge of nose propertes Medcne s full of estmaton problems IDEA Lab, Radology, Cornell 21

22 Dfferent approaches to estmaton Mnmum varance unbased estmators Least Squares Maxmum-lkelhood Maxmum entropy Maxmum a posteror uses pror nformaton about θ has no statstcal bass uses knowledge of nose PDF IDEA Lab, Radology, Cornell 22

23 Probablty vs. Statstcs Probablty: Mathematcal models of uncertanty predct outcomes Ths s the heart of probablty Models, and ther consequences What s the probablty of a model generatng some partcular data as an outcome? Statstcs: Gven an outcome, analyze dfferent models Dd ths model generate the data? From among dfferent models (or parameters), whch one generated the data? IDEA Lab, Radology, Cornell 23

24 Maxmul Lkelhood Estmator for Lne Fttng Problem What f we know somethng about the nose?.e. Pr(n) If nose not unform across samples, LS mght be ncorrect nose σ = 10 E( y nt Best ( y y h( x, m) = ( σ nt, m) ) ) 2 = arg mn E( y, nt where, m) h( x) = y nt + mx Ths can be wrtten n matrx notaton, as θ ML = arg mn W(y Hθ) 2 nose σ = 0.1 ML estmate What s W? IDEA Lab, Radology, Cornell 24

25 Defnton of lkelhood Lkelhood s a probablty model of the uncertanty n output gven a known nput The lkelhood of a hypothess s the probablty that t would have resulted n the data you saw Thnk of the data as fxed, and try to chose among the possble PDF s Often, a parameterzed famly of PDF s ML parameter estmaton IDEA Lab, Radology, Cornell 25

26 Gaussan Nose Models In lnear model we dscussed, lkelhood comes from nose statstcs Smple dea: want to ncorporate knowledge of nose statstcs If unform whte Gaussan nose: 2 n 1 n = 1 Pr( n) exp = exp 2 2 Z 2σ Z 2σ If non-unform whte Gaussan nose: Pr( n) = 1 exp Z n 2σ IDEA Lab, Radology, Cornell 26

27 Maxmum Lkelhood Estmator - Theory n = y-hθ, Pr(n) = exp(- n 2 /2σ 2 ) Therefore Pr(y for known θ) = Pr(n) Smple dea: want to maxmze Pr(y θ) - called the lkelhood functon Example 1: show that for unform ndependent Gaussan nose θ ML = arg mn y-hθ 2 Example 2: For non-unform Gaussan nose θ ML = arg mn W(y-Hθ) 2 IDEA Lab, Radology, Cornell 27

28 MLE Bottomlne: Use nose propertes to wrte Pr(y θ) Whchever θ maxmze above, s the MLE IDEA Lab, Radology, Cornell 28

29 Example Estmatng man frequency of ECG sgnal Model: y(t ) = a sn(f t ) + n What s the MLE of a, f? Pr(y θ ) = exp(-σ (y(t ) - a sn(f t ) ) 2 / 2 σ 2 ) IDEA Lab, Radology, Cornell 29

30 Maxmum Lkelhood Detecton ML s qute a general, powerful dea Same deas can be used for classfcaton and detecton of features hdden n data Example 1: Decdng whether a voxel s artery or ven There are 3 hypotheses at each voxel: Voxel s artery, or voxel s ven, or voxel s parenchyma IDEA Lab, Radology, Cornell 30

31 Example: MRA segmentaton artery/ven may have smlar ntensty at gven tme pont but dfferent tme profles wsh to segment accordng to tme profle, not sngle ntensty IDEA Lab, Radology, Cornell 31

32 Expected Result IDEA Lab, Radology, Cornell 32

33 Example: MRA segmentaton Frst: need a tme model of all segments h( t θa ) h ( v t θ ) Lets use ML prncples to see whch voxel belongs to whch model Artery: Ven: Parench: y = h( t θa ) + n y = h( t θv ) + n y = h( t θ ) + n p IDEA Lab, Radology, Cornell

34 Maxmum Lkelhood Classfcaton Artery: Ven: Paren: y = h( t θ ) + n y = h( t θ ) + n y = h( t θ ) + n a v p Pr( y θ ) IDEA Lab, Radology, Cornell Pr( y Pr( y a θ ) v θ ) p ( y exp ( y exp ( y exp = 2 = 2 h( t θa )) 2σ h( t θv )) 2σ h( t θ )) = 2 So at each voxel, the best model s one that maxmzes: 2 ( y h( t Pr( y = θ )) θ ) Pr( y θ ) = exp 2 2σ Or equvalently, mnmzes: ( y h( t θ )) 2 2σ p 2 2 2

35 Lver tumour quantfcaton from Dynamc Contrast Enhanced MRI Data: Tumor model Rabbt DCE-MR data Paramegnetc contrast agent, pathology gold standard Extract temporal features from DCE-MRI Use these features for accurate detecton and quantfcaton of tumour IDEA Lab, Radology, Cornell

36 Lver Tumour Temporal models Temporal models used to extract features Typcal plot of tme-resolved MR sgnal of varous tssue classes h( t θ ) IDEA Lab, Radology, Cornell 36

37 Lver tumour quantfcaton from DCE-MRI IDEA Lab, Radology, Cornell

38 ML Tutorals Sldes by Andrew Moore (CMU): avalable on course webpage Paper by Jae Myung, Tutoral on Maxmum Lkelhood : avalable on course webpage Max Entropy Tutoral IDEA Lab, Radology, Cornell 38

39 Next lecture (Frday) Non-parametrc densty estmaton Hstograms + varous fttng methods Nearest neghbor Parzen estmaton Next lecture (Wednesday) Maxmum A Posteror Estmators Several examples IDEA Lab, Radology, Cornell 39

40 CS5540: Computatonal Technques for Analyzng Clncal Data Lecture 7: Statstcal Estmaton: Least Squares, Maxmum Lkelhood and Maxmum A Posteror Estmators Ashsh Raj, PhD Image Data Evaluaton and Analytcs Laboratory (IDEAL) Department of Radology Well Cornell Medcal College New York

41 Falure modes of ML Lkelhood sn t the only crteron for selectng a model or parameter Though t s obvously an mportant one Bzarre models may have hgh lkelhood Consder a speedometer readng 55 MPH Lkelhood of true speed = 55 : 10% Lkelhood of speedometer stuck : 100% ML lkes fary tales In practce, exclude such hypotheses There must be a prncpled soluton IDEA Lab, Radology, Cornell 41

42 Maxmum a Posteror Estmate Ths s an example of usng an mage pror Prors are generally expressed n the form of a PDF Pr(x) Once the lkelhood L(x) and pror are known, we have complete statstcal knowledge LS/ML are suboptmal n presence of pror MAP (aka Bayesan) estmates are optmal lkelhood Bayes Theorem: posteror Pr(x y) = Pr(y x). Pr(x) pror Pr(y) IDEA Lab, Radology, Cornell 42

43 Maxmum a Posteror (Bayesan) Estmate Consder the class of lnear systems y = Hx + n Bayesan methods maxmze the posteror probablty: Pr(x y) Pr(y x). Pr(x) Pr(y x) (lkelhood functon) = exp(- y-hx 2 ) Pr(x) (pror PDF) = exp(-g(x)) Non-Bayesan: maxmze only lkelhood x est = arg mn y-hx 2 Bayesan: x est = arg mn y-hx 2 + G(x), where G(x) s obtaned from the pror dstrbuton of x If G(x) = Gx 2 Tkhonov Regularzaton IDEA Lab, Radology, Cornell 43

44 Other example of Estmaton n MR Image denosng: H = I Image deblurrng: H = convoluton mtx n mg-space Super-resoluton: H = dagonal mtx n k-space Metabolte quantfcaton n MRSI IDEA Lab, Radology, Cornell 44

45 What Is the Rght Imagng Model? y = H x + n, n s Gaussan (1) MAP Sense y = H x + n, n, x are Gaussan (2) MAP Sense = Bayesan (MAP) estmate of (2) IDEA Lab, Radology, Cornell 45

46 Intro to Bayesan Estmaton Bayesan methods maxmze the posteror probablty: Pr(x y) Pr(y x). Pr(x) Pr(y x) (lkelhood functon) = exp(- y-hx 2 ) Pr(x) (pror PDF) = exp(-g(x)) Gaussan pror: Pr(x) = exp{- ½ x T R x -1 x} MAP estmate: x est = arg mn y-hx 2 + G(x) MAP estmate for Gaussan everythng s known as Wener estmate IDEA Lab, Radology, Cornell 46

47 Regularzaton = Bayesan Estmaton! For any regularzaton scheme, ts almost always possble to formulate the correspondng MAP problem MAP = superset of regularzaton Pror model MAP Regularzaton scheme So why deal wth regularzaton?? IDEA Lab, Radology, Cornell 47

48 Lets talk about Pror Models Temporal prors: smooth tme-trajectory Sparse prors: L0, L1, L2 (=Tkhonov) Spatal Prors: most powerful for mages I recommend robust spatal prors usng Markov Felds Want prors to be general, not too specfc Ie, weak rather than strong prors IDEA Lab, Radology, Cornell 48

49 How to do regularzaton Frst model physcal property of mage, then create a pror whch captures t, then formulate MAP estmator, Then fnd a good algorthm to solve t! How NOT to do regularzaton DON T use regularzaton scheme wthout bearng on physcal property of mage! Example: L1 or L0 pror n k-space! Specfcally: deblurrng n k-space (handy b/c convoluton becomes multply) BUT: hard to mpose smoothness prors n k-space no meanng IDEA Lab, Radology, Cornell 49

50 MAP for Lne fttng problem If model estmated by ML and Pror nfo do not agree MAP s a compromse between the two LS estmate MAP estmate Most probable pror model IDEA Lab, Radology, Cornell 50

51 Mult-varate FLASH Acqure 6-10 accelerated FLASH data sets at dfferent flp angles or TR s Generate T 1 maps by fttng to: S ( ) ( * ) 1 exp TR T = exp TE T snα 2 1 cos exp α Not enough nfo n a sngle voxel Nose causes ncorrect estmates Error n flp angle vares spatally! 1 ( TR T ) 1 IDEA Lab, Radology, Cornell 51

52 Spatally Coherent T 1, ρ estmaton Frst, stack parameters from all voxels n one bg vector x Stack all observed flp angle mages n y Then we can wrte y = M (x) + n Recall M s the (nonlnear) functonal obtaned from S ( 1) ( TR T ) ( * 1 exp TR T = exp TE T ) 2 snα 1 cos α exp Solve for x by non-lnear least square fttng, PLUS spatal pror: x est = arg mn x y - M (x) 2 + µ 2 Dx 2 E(x) Makes M(x) close to y Makes x smooth Mnmze va MATLAB s lsqnonln functon How? Construct δ = [y - M (x); µ Dx]. Then E(x) = δ 2 1 IDEA Lab, Radology, Cornell 52

53 Mult-Flp Results combned ρ, T 1 n pseudocolour IDEA Lab, Radology, Cornell 53

54 Mult-Flp Results combned ρ, T 1 n pseudocolour IDEA Lab, Radology, Cornell 54

55 Spatal Prors For Images - Example Frames are tghtly dstrbuted around mean After subtractng mean, mages are close to Gaussan varance mean tme frame N f frame 1 frame 2 mean µ x (,j) envelope a(,j) Pror: -mean s µ x -local std.dev. vares as a(,j) IDEA Lab, Radology, Cornell 55

56 Spatal Prors for MR mages statonary Stochastc MR mage model: process x(,j) = µ x (,j) + a(,j). (h ** p)(,j) (1) ** denotes 2D convoluton µ x (,j) s mean mage for class p(,j) s a unt varance..d. stochastc process a(,j) s an envelope functon h(,j) smulates correlaton propertes of mage x r(τ 1, τ 2 ) = (h ** h)(τ 1, τ 2 ) x = ACp + µ (2) where A = dag(a), and C s the Toepltz matrx generated by h Can model many mportant statonary and non-statonary cases IDEA Lab, Radology, Cornell 56

57 MAP estmate for Imagng Model (3) The Wener estmate x MAP - µ x = HR x (HR x H H + R n ) -1 (y- µ y ) (3) R x, R n = covarance matrces of x and n Statonarty R x has Toepltz structure fast processng x MAP - µ x = HACC H A H ( HACC H A H H H + σ n 2 I ) -1 (y- µ y ) (4) Drect nverson prohbtve; so use CG teratve method (4) better than (3) snce A and C are O(N log N) operatons, enablng much faster processng IDEA Lab, Radology, Cornell 57

58 How to obtan estmates of A, C? Need a tranng set of full-resoluton mages x k, k = 1,,K Parallel magng doesnt provde un-alased full-res mages Approaches: 1. Use prevous full-res scans - tme consumng, need frequent updatng 2. Use SENSE-reconstructed mages for tranng set - very bad nose amplfcaton ssues for hgh speedu ps 3. Drectly estmate parameters from avalable parallel data - Alasng may cause naccuraces IDEA Lab, Radology, Cornell 58

59 MAP for Dynamc Parallel Imagng N f mages avalable for parameter estmaton! All mages well-regstered Tghtly dstrbuted around pxel mean Parameters can be estmated from alased data tme frame N f frame 2 frame 1 k y frame 1 frame 2 frame N f N f /R full-res mages k x IDEA Lab, Radology, Cornell 59

60 MAP-SENSE Prelmnary Results Scans acceleraty 5x The angogram was computed by: avg(post-contrast) avg(pre-contrast) Unaccelerated 5x faster: SENSE 5x faster: MAP-SENSE IDEA Lab, Radology, Cornell 60

61 References Smon Kay. Statstcal Sgnal Processng. Part I: Estmaton Theory. Prentce Hall 2002 Smon Kay. Statstcal Sgnal Processng. Part II: Detecton Theory. Prentce Hall 2002 Haacke et al. Fundamentals of MRI. Zh-Pe Lang and Paul Lauterbur. Prncples of MRI A Sgnal Processng Perspectve. Info on part IV: Ashsh Raj. Improvements n MRI Usng Informaton Redundancy. PhD thess, Cornell Unversty, May Webste: IDEA Lab, Radology, Cornell 61

62 Maxmum Lkelhood Estmator But f nose s jontly Gaussan wth cov. matrx C Recall C, E(nn T ). Then Pr(n) = e -½ nt C -1 n L(y θ) = ½ (y-hθ) T C -1 (y-hθ) θ ML = argmn ½ (y-hθ) T C -1 (y-hθ) Ths also has a closed form soluton θ ML = (H T C -1 H) -1 H T C -1 y If n s not Gaussan at all, ML estmators become complcated and non-lnear Fortunately, n MR nose s usually Gaussan IDEA Lab, Radology, Cornell 62

2E Pattern Recognition Solutions to Introduction to Pattern Recognition, Chapter 2: Bayesian pattern classification

2E Pattern Recognition Solutions to Introduction to Pattern Recognition, Chapter 2: Bayesian pattern classification E395 - Pattern Recognton Solutons to Introducton to Pattern Recognton, Chapter : Bayesan pattern classfcaton Preface Ths document s a soluton manual for selected exercses from Introducton to Pattern Recognton

More information

Lecture Notes on Linear Regression

Lecture Notes on Linear Regression Lecture Notes on Lnear Regresson Feng L fl@sdueducn Shandong Unversty, Chna Lnear Regresson Problem In regresson problem, we am at predct a contnuous target value gven an nput feature vector We assume

More information

Composite Hypotheses testing

Composite Hypotheses testing Composte ypotheses testng In many hypothess testng problems there are many possble dstrbutons that can occur under each of the hypotheses. The output of the source s a set of parameters (ponts n a parameter

More information

Maximum Likelihood Estimation (MLE)

Maximum Likelihood Estimation (MLE) Maxmum Lkelhood Estmaton (MLE) Ken Kreutz-Delgado (Nuno Vasconcelos) ECE 175A Wnter 01 UCSD Statstcal Learnng Goal: Gven a relatonshp between a feature vector x and a vector y, and d data samples (x,y

More information

P R. Lecture 4. Theory and Applications of Pattern Recognition. Dept. of Electrical and Computer Engineering /

P R. Lecture 4. Theory and Applications of Pattern Recognition. Dept. of Electrical and Computer Engineering / Theory and Applcatons of Pattern Recognton 003, Rob Polkar, Rowan Unversty, Glassboro, NJ Lecture 4 Bayes Classfcaton Rule Dept. of Electrcal and Computer Engneerng 0909.40.0 / 0909.504.04 Theory & Applcatons

More information

MLE and Bayesian Estimation. Jie Tang Department of Computer Science & Technology Tsinghua University 2012

MLE and Bayesian Estimation. Jie Tang Department of Computer Science & Technology Tsinghua University 2012 MLE and Bayesan Estmaton Je Tang Department of Computer Scence & Technology Tsnghua Unversty 01 1 Lnear Regresson? As the frst step, we need to decde how we re gong to represent the functon f. One example:

More information

ENG 8801/ Special Topics in Computer Engineering: Pattern Recognition. Memorial University of Newfoundland Pattern Recognition

ENG 8801/ Special Topics in Computer Engineering: Pattern Recognition. Memorial University of Newfoundland Pattern Recognition EG 880/988 - Specal opcs n Computer Engneerng: Pattern Recognton Memoral Unversty of ewfoundland Pattern Recognton Lecture 7 May 3, 006 http://wwwengrmunca/~charlesr Offce Hours: uesdays hursdays 8:30-9:30

More information

Lecture 10 Support Vector Machines II

Lecture 10 Support Vector Machines II Lecture 10 Support Vector Machnes II 22 February 2016 Taylor B. Arnold Yale Statstcs STAT 365/665 1/28 Notes: Problem 3 s posted and due ths upcomng Frday There was an early bug n the fake-test data; fxed

More information

CS 2750 Machine Learning. Lecture 5. Density estimation. CS 2750 Machine Learning. Announcements

CS 2750 Machine Learning. Lecture 5. Density estimation. CS 2750 Machine Learning. Announcements CS 750 Machne Learnng Lecture 5 Densty estmaton Mlos Hauskrecht mlos@cs.ptt.edu 539 Sennott Square CS 750 Machne Learnng Announcements Homework Due on Wednesday before the class Reports: hand n before

More information

8/25/17. Data Modeling. Data Modeling. Data Modeling. Patrice Koehl Department of Biological Sciences National University of Singapore

8/25/17. Data Modeling. Data Modeling. Data Modeling. Patrice Koehl Department of Biological Sciences National University of Singapore 8/5/17 Data Modelng Patrce Koehl Department of Bologcal Scences atonal Unversty of Sngapore http://www.cs.ucdavs.edu/~koehl/teachng/bl59 koehl@cs.ucdavs.edu Data Modelng Ø Data Modelng: least squares Ø

More information

Lecture 12: Classification

Lecture 12: Classification Lecture : Classfcaton g Dscrmnant functons g The optmal Bayes classfer g Quadratc classfers g Eucldean and Mahalanobs metrcs g K Nearest Neghbor Classfers Intellgent Sensor Systems Rcardo Guterrez-Osuna

More information

Motion Perception Under Uncertainty. Hongjing Lu Department of Psychology University of Hong Kong

Motion Perception Under Uncertainty. Hongjing Lu Department of Psychology University of Hong Kong Moton Percepton Under Uncertanty Hongjng Lu Department of Psychology Unversty of Hong Kong Outlne Uncertanty n moton stmulus Correspondence problem Qualtatve fttng usng deal observer models Based on sgnal

More information

Linear Approximation with Regularization and Moving Least Squares

Linear Approximation with Regularization and Moving Least Squares Lnear Approxmaton wth Regularzaton and Movng Least Squares Igor Grešovn May 007 Revson 4.6 (Revson : March 004). 5 4 3 0.5 3 3.5 4 Contents: Lnear Fttng...4. Weghted Least Squares n Functon Approxmaton...

More information

Uncertainty as the Overlap of Alternate Conditional Distributions

Uncertainty as the Overlap of Alternate Conditional Distributions Uncertanty as the Overlap of Alternate Condtonal Dstrbutons Olena Babak and Clayton V. Deutsch Centre for Computatonal Geostatstcs Department of Cvl & Envronmental Engneerng Unversty of Alberta An mportant

More information

Generalized Linear Methods

Generalized Linear Methods Generalzed Lnear Methods 1 Introducton In the Ensemble Methods the general dea s that usng a combnaton of several weak learner one could make a better learner. More formally, assume that we have a set

More information

Clustering & Unsupervised Learning

Clustering & Unsupervised Learning Clusterng & Unsupervsed Learnng Ken Kreutz-Delgado (Nuno Vasconcelos) ECE 175A Wnter 2012 UCSD Statstcal Learnng Goal: Gven a relatonshp between a feature vector x and a vector y, and d data samples (x,y

More information

The conjugate prior to a Bernoulli is. A) Bernoulli B) Gaussian C) Beta D) none of the above

The conjugate prior to a Bernoulli is. A) Bernoulli B) Gaussian C) Beta D) none of the above The conjugate pror to a Bernoull s A) Bernoull B) Gaussan C) Beta D) none of the above The conjugate pror to a Gaussan s A) Bernoull B) Gaussan C) Beta D) none of the above MAP estmates A) argmax θ p(θ

More information

Tutorial 2. COMP4134 Biometrics Authentication. February 9, Jun Xu, Teaching Asistant

Tutorial 2. COMP4134 Biometrics Authentication. February 9, Jun Xu, Teaching Asistant Tutoral 2 COMP434 ometrcs uthentcaton Jun Xu, Teachng sstant csjunxu@comp.polyu.edu.hk February 9, 207 Table of Contents Problems Problem : nswer the questons Problem 2: Power law functon Problem 3: Convoluton

More information

Global Sensitivity. Tuesday 20 th February, 2018

Global Sensitivity. Tuesday 20 th February, 2018 Global Senstvty Tuesday 2 th February, 28 ) Local Senstvty Most senstvty analyses [] are based on local estmates of senstvty, typcally by expandng the response n a Taylor seres about some specfc values

More information

Tracking with Kalman Filter

Tracking with Kalman Filter Trackng wth Kalman Flter Scott T. Acton Vrgna Image and Vdeo Analyss (VIVA), Charles L. Brown Department of Electrcal and Computer Engneerng Department of Bomedcal Engneerng Unversty of Vrgna, Charlottesvlle,

More information

Statistical pattern recognition

Statistical pattern recognition Statstcal pattern recognton Bayes theorem Problem: decdng f a patent has a partcular condton based on a partcular test However, the test s mperfect Someone wth the condton may go undetected (false negatve

More information

Conjugacy and the Exponential Family

Conjugacy and the Exponential Family CS281B/Stat241B: Advanced Topcs n Learnng & Decson Makng Conjugacy and the Exponental Famly Lecturer: Mchael I. Jordan Scrbes: Bran Mlch 1 Conjugacy In the prevous lecture, we saw conjugate prors for the

More information

Linear Feature Engineering 11

Linear Feature Engineering 11 Lnear Feature Engneerng 11 2 Least-Squares 2.1 Smple least-squares Consder the followng dataset. We have a bunch of nputs x and correspondng outputs y. The partcular values n ths dataset are x y 0.23 0.19

More information

Lossy Compression. Compromise accuracy of reconstruction for increased compression.

Lossy Compression. Compromise accuracy of reconstruction for increased compression. Lossy Compresson Compromse accuracy of reconstructon for ncreased compresson. The reconstructon s usually vsbly ndstngushable from the orgnal mage. Typcally, one can get up to 0:1 compresson wth almost

More information

Homework Assignment 3 Due in class, Thursday October 15

Homework Assignment 3 Due in class, Thursday October 15 Homework Assgnment 3 Due n class, Thursday October 15 SDS 383C Statstcal Modelng I 1 Rdge regresson and Lasso 1. Get the Prostrate cancer data from http://statweb.stanford.edu/~tbs/elemstatlearn/ datasets/prostate.data.

More information

Support Vector Machines. Vibhav Gogate The University of Texas at dallas

Support Vector Machines. Vibhav Gogate The University of Texas at dallas Support Vector Machnes Vbhav Gogate he Unversty of exas at dallas What We have Learned So Far? 1. Decson rees. Naïve Bayes 3. Lnear Regresson 4. Logstc Regresson 5. Perceptron 6. Neural networks 7. K-Nearest

More information

Lectures - Week 4 Matrix norms, Conditioning, Vector Spaces, Linear Independence, Spanning sets and Basis, Null space and Range of a Matrix

Lectures - Week 4 Matrix norms, Conditioning, Vector Spaces, Linear Independence, Spanning sets and Basis, Null space and Range of a Matrix Lectures - Week 4 Matrx norms, Condtonng, Vector Spaces, Lnear Independence, Spannng sets and Bass, Null space and Range of a Matrx Matrx Norms Now we turn to assocatng a number to each matrx. We could

More information

Automatic Object Trajectory- Based Motion Recognition Using Gaussian Mixture Models

Automatic Object Trajectory- Based Motion Recognition Using Gaussian Mixture Models Automatc Object Trajectory- Based Moton Recognton Usng Gaussan Mxture Models Fasal I. Bashr, Ashfaq A. Khokhar, Dan Schonfeld Electrcal and Computer Engneerng, Unversty of Illnos at Chcago. Chcago, IL,

More information

Chapter 11: Simple Linear Regression and Correlation

Chapter 11: Simple Linear Regression and Correlation Chapter 11: Smple Lnear Regresson and Correlaton 11-1 Emprcal Models 11-2 Smple Lnear Regresson 11-3 Propertes of the Least Squares Estmators 11-4 Hypothess Test n Smple Lnear Regresson 11-4.1 Use of t-tests

More information

Lecture 3: Shannon s Theorem

Lecture 3: Shannon s Theorem CSE 533: Error-Correctng Codes (Autumn 006 Lecture 3: Shannon s Theorem October 9, 006 Lecturer: Venkatesan Guruswam Scrbe: Wdad Machmouch 1 Communcaton Model The communcaton model we are usng conssts

More information

Lecture 3. Ax x i a i. i i

Lecture 3. Ax x i a i. i i 18.409 The Behavor of Algorthms n Practce 2/14/2 Lecturer: Dan Spelman Lecture 3 Scrbe: Arvnd Sankar 1 Largest sngular value In order to bound the condton number, we need an upper bound on the largest

More information

Module 3 LOSSY IMAGE COMPRESSION SYSTEMS. Version 2 ECE IIT, Kharagpur

Module 3 LOSSY IMAGE COMPRESSION SYSTEMS. Version 2 ECE IIT, Kharagpur Module 3 LOSSY IMAGE COMPRESSION SYSTEMS Verson ECE IIT, Kharagpur Lesson 6 Theory of Quantzaton Verson ECE IIT, Kharagpur Instructonal Objectves At the end of ths lesson, the students should be able to:

More information

Week 5: Neural Networks

Week 5: Neural Networks Week 5: Neural Networks Instructor: Sergey Levne Neural Networks Summary In the prevous lecture, we saw how we can construct neural networks by extendng logstc regresson. Neural networks consst of multple

More information

Hidden Markov Models

Hidden Markov Models Hdden Markov Models Namrata Vaswan, Iowa State Unversty Aprl 24, 204 Hdden Markov Model Defntons and Examples Defntons:. A hdden Markov model (HMM) refers to a set of hdden states X 0, X,..., X t,...,

More information

Linear Classification, SVMs and Nearest Neighbors

Linear Classification, SVMs and Nearest Neighbors 1 CSE 473 Lecture 25 (Chapter 18) Lnear Classfcaton, SVMs and Nearest Neghbors CSE AI faculty + Chrs Bshop, Dan Klen, Stuart Russell, Andrew Moore Motvaton: Face Detecton How do we buld a classfer to dstngush

More information

Econ107 Applied Econometrics Topic 3: Classical Model (Studenmund, Chapter 4)

Econ107 Applied Econometrics Topic 3: Classical Model (Studenmund, Chapter 4) I. Classcal Assumptons Econ7 Appled Econometrcs Topc 3: Classcal Model (Studenmund, Chapter 4) We have defned OLS and studed some algebrac propertes of OLS. In ths topc we wll study statstcal propertes

More information

Clustering & (Ken Kreutz-Delgado) UCSD

Clustering & (Ken Kreutz-Delgado) UCSD Clusterng & Unsupervsed Learnng Nuno Vasconcelos (Ken Kreutz-Delgado) UCSD Statstcal Learnng Goal: Gven a relatonshp between a feature vector x and a vector y, and d data samples (x,y ), fnd an approxmatng

More information

Finite Mixture Models and Expectation Maximization. Most slides are from: Dr. Mario Figueiredo, Dr. Anil Jain and Dr. Rong Jin

Finite Mixture Models and Expectation Maximization. Most slides are from: Dr. Mario Figueiredo, Dr. Anil Jain and Dr. Rong Jin Fnte Mxture Models and Expectaton Maxmzaton Most sldes are from: Dr. Maro Fgueredo, Dr. Anl Jan and Dr. Rong Jn Recall: The Supervsed Learnng Problem Gven a set of n samples X {(x, y )},,,n Chapter 3 of

More information

Thermodynamics and statistical mechanics in materials modelling II

Thermodynamics and statistical mechanics in materials modelling II Course MP3 Lecture 8/11/006 (JAE) Course MP3 Lecture 8/11/006 Thermodynamcs and statstcal mechancs n materals modellng II A bref résumé of the physcal concepts used n materals modellng Dr James Ellott.1

More information

CIS526: Machine Learning Lecture 3 (Sept 16, 2003) Linear Regression. Preparation help: Xiaoying Huang. x 1 θ 1 output... θ M x M

CIS526: Machine Learning Lecture 3 (Sept 16, 2003) Linear Regression. Preparation help: Xiaoying Huang. x 1 θ 1 output... θ M x M CIS56: achne Learnng Lecture 3 (Sept 6, 003) Preparaton help: Xaoyng Huang Lnear Regresson Lnear regresson can be represented by a functonal form: f(; θ) = θ 0 0 +θ + + θ = θ = 0 ote: 0 s a dummy attrbute

More information

The Gaussian classifier. Nuno Vasconcelos ECE Department, UCSD

The Gaussian classifier. Nuno Vasconcelos ECE Department, UCSD he Gaussan classfer Nuno Vasconcelos ECE Department, UCSD Bayesan decson theory recall that we have state of the world X observatons g decson functon L[g,y] loss of predctng y wth g Bayes decson rule s

More information

xp(x µ) = 0 p(x = 0 µ) + 1 p(x = 1 µ) = µ

xp(x µ) = 0 p(x = 0 µ) + 1 p(x = 1 µ) = µ CSE 455/555 Sprng 2013 Homework 7: Parametrc Technques Jason J. Corso Computer Scence and Engneerng SUY at Buffalo jcorso@buffalo.edu Solutons by Yngbo Zhou Ths assgnment does not need to be submtted and

More information

The big picture. Outline

The big picture. Outline The bg pcture Vncent Claveau IRISA - CNRS, sldes from E. Kjak INSA Rennes Notatons classes: C = {ω = 1,.., C} tranng set S of sze m, composed of m ponts (x, ω ) per class ω representaton space: R d (=

More information

Change Detection: Current State of the Art and Future Directions

Change Detection: Current State of the Art and Future Directions Change Detecton: Current State of the Art and Future Drectons Dapeng Olver Wu Electrcal & Computer Engneerng Unversty of Florda http://www.wu.ece.ufl.edu/ Outlne Motvaton & problem statement Change detecton

More information

Chapter 2 - The Simple Linear Regression Model S =0. e i is a random error. S β2 β. This is a minimization problem. Solution is a calculus exercise.

Chapter 2 - The Simple Linear Regression Model S =0. e i is a random error. S β2 β. This is a minimization problem. Solution is a calculus exercise. Chapter - The Smple Lnear Regresson Model The lnear regresson equaton s: where y + = β + β e for =,..., y and are observable varables e s a random error How can an estmaton rule be constructed for the

More information

Communication with AWGN Interference

Communication with AWGN Interference Communcaton wth AWG Interference m {m } {p(m } Modulator s {s } r=s+n Recever ˆm AWG n m s a dscrete random varable(rv whch takes m wth probablty p(m. Modulator maps each m nto a waveform sgnal s m=m

More information

Maximum Likelihood Estimation of Binary Dependent Variables Models: Probit and Logit. 1. General Formulation of Binary Dependent Variables Models

Maximum Likelihood Estimation of Binary Dependent Variables Models: Probit and Logit. 1. General Formulation of Binary Dependent Variables Models ECO 452 -- OE 4: Probt and Logt Models ECO 452 -- OE 4 Maxmum Lkelhood Estmaton of Bnary Dependent Varables Models: Probt and Logt hs note demonstrates how to formulate bnary dependent varables models

More information

Stat 543 Exam 2 Spring 2016

Stat 543 Exam 2 Spring 2016 Stat 543 Exam 2 Sprng 206 I have nether gven nor receved unauthorzed assstance on ths exam. Name Sgned Date Name Prnted Ths Exam conssts of questons. Do at least 0 of the parts of the man exam. I wll score

More information

Using T.O.M to Estimate Parameter of distributions that have not Single Exponential Family

Using T.O.M to Estimate Parameter of distributions that have not Single Exponential Family IOSR Journal of Mathematcs IOSR-JM) ISSN: 2278-5728. Volume 3, Issue 3 Sep-Oct. 202), PP 44-48 www.osrjournals.org Usng T.O.M to Estmate Parameter of dstrbutons that have not Sngle Exponental Famly Jubran

More information

For now, let us focus on a specific model of neurons. These are simplified from reality but can achieve remarkable results.

For now, let us focus on a specific model of neurons. These are simplified from reality but can achieve remarkable results. Neural Networks : Dervaton compled by Alvn Wan from Professor Jtendra Malk s lecture Ths type of computaton s called deep learnng and s the most popular method for many problems, such as computer vson

More information

SIO 224. m(r) =(ρ(r),k s (r),µ(r))

SIO 224. m(r) =(ρ(r),k s (r),µ(r)) SIO 224 1. A bref look at resoluton analyss Here s some background for the Masters and Gubbns resoluton paper. Global Earth models are usually found teratvely by assumng a startng model and fndng small

More information

Chat eld, C. and A.J.Collins, Introduction to multivariate analysis. Chapman & Hall, 1980

Chat eld, C. and A.J.Collins, Introduction to multivariate analysis. Chapman & Hall, 1980 MT07: Multvarate Statstcal Methods Mke Tso: emal mke.tso@manchester.ac.uk Webpage for notes: http://www.maths.manchester.ac.uk/~mkt/new_teachng.htm. Introducton to multvarate data. Books Chat eld, C. and

More information

MATH 829: Introduction to Data Mining and Analysis The EM algorithm (part 2)

MATH 829: Introduction to Data Mining and Analysis The EM algorithm (part 2) 1/16 MATH 829: Introducton to Data Mnng and Analyss The EM algorthm (part 2) Domnque Gullot Departments of Mathematcal Scences Unversty of Delaware Aprl 20, 2016 Recall 2/16 We are gven ndependent observatons

More information

Limited Dependent Variables

Limited Dependent Variables Lmted Dependent Varables. What f the left-hand sde varable s not a contnuous thng spread from mnus nfnty to plus nfnty? That s, gven a model = f (, β, ε, where a. s bounded below at zero, such as wages

More information

Markov Chain Monte Carlo (MCMC), Gibbs Sampling, Metropolis Algorithms, and Simulated Annealing Bioinformatics Course Supplement

Markov Chain Monte Carlo (MCMC), Gibbs Sampling, Metropolis Algorithms, and Simulated Annealing Bioinformatics Course Supplement Markov Chan Monte Carlo MCMC, Gbbs Samplng, Metropols Algorthms, and Smulated Annealng 2001 Bonformatcs Course Supplement SNU Bontellgence Lab http://bsnuackr/ Outlne! Markov Chan Monte Carlo MCMC! Metropols-Hastngs

More information

Space of ML Problems. CSE 473: Artificial Intelligence. Parameter Estimation and Bayesian Networks. Learning Topics

Space of ML Problems. CSE 473: Artificial Intelligence. Parameter Estimation and Bayesian Networks. Learning Topics /7/7 CSE 73: Artfcal Intellgence Bayesan - Learnng Deter Fox Sldes adapted from Dan Weld, Jack Breese, Dan Klen, Daphne Koller, Stuart Russell, Andrew Moore & Luke Zettlemoyer What s Beng Learned? Space

More information

Machine learning: Density estimation

Machine learning: Density estimation CS 70 Foundatons of AI Lecture 3 Machne learnng: ensty estmaton Mlos Hauskrecht mlos@cs.ptt.edu 539 Sennott Square ata: ensty estmaton {.. n} x a vector of attrbute values Objectve: estmate the model of

More information

Stat 543 Exam 2 Spring 2016

Stat 543 Exam 2 Spring 2016 Stat 543 Exam 2 Sprng 2016 I have nether gven nor receved unauthorzed assstance on ths exam. Name Sgned Date Name Prnted Ths Exam conssts of 11 questons. Do at least 10 of the 11 parts of the man exam.

More information

since [1-( 0+ 1x1i+ 2x2 i)] [ 0+ 1x1i+ assumed to be a reasonable approximation

since [1-( 0+ 1x1i+ 2x2 i)] [ 0+ 1x1i+ assumed to be a reasonable approximation Econ 388 R. Butler 204 revsons Lecture 4 Dummy Dependent Varables I. Lnear Probablty Model: the Regresson model wth a dummy varables as the dependent varable assumpton, mplcaton regular multple regresson

More information

Dr. Shalabh Department of Mathematics and Statistics Indian Institute of Technology Kanpur

Dr. Shalabh Department of Mathematics and Statistics Indian Institute of Technology Kanpur Analyss of Varance and Desgn of Exerments-I MODULE III LECTURE - 2 EXPERIMENTAL DESIGN MODELS Dr. Shalabh Deartment of Mathematcs and Statstcs Indan Insttute of Technology Kanur 2 We consder the models

More information

Comparison of Regression Lines

Comparison of Regression Lines STATGRAPHICS Rev. 9/13/2013 Comparson of Regresson Lnes Summary... 1 Data Input... 3 Analyss Summary... 4 Plot of Ftted Model... 6 Condtonal Sums of Squares... 6 Analyss Optons... 7 Forecasts... 8 Confdence

More information

Instance-Based Learning (a.k.a. memory-based learning) Part I: Nearest Neighbor Classification

Instance-Based Learning (a.k.a. memory-based learning) Part I: Nearest Neighbor Classification Instance-Based earnng (a.k.a. memory-based learnng) Part I: Nearest Neghbor Classfcaton Note to other teachers and users of these sldes. Andrew would be delghted f you found ths source materal useful n

More information

CSci 6974 and ECSE 6966 Math. Tech. for Vision, Graphics and Robotics Lecture 21, April 17, 2006 Estimating A Plane Homography

CSci 6974 and ECSE 6966 Math. Tech. for Vision, Graphics and Robotics Lecture 21, April 17, 2006 Estimating A Plane Homography CSc 6974 and ECSE 6966 Math. Tech. for Vson, Graphcs and Robotcs Lecture 21, Aprl 17, 2006 Estmatng A Plane Homography Overvew We contnue wth a dscusson of the major ssues, usng estmaton of plane projectve

More information

Why Bayesian? 3. Bayes and Normal Models. State of nature: class. Decision rule. Rev. Thomas Bayes ( ) Bayes Theorem (yes, the famous one)

Why Bayesian? 3. Bayes and Normal Models. State of nature: class. Decision rule. Rev. Thomas Bayes ( ) Bayes Theorem (yes, the famous one) Why Bayesan? 3. Bayes and Normal Models Alex M. Martnez alex@ece.osu.edu Handouts Handoutsfor forece ECE874 874Sp Sp007 If all our research (n PR was to dsappear and you could only save one theory, whch

More information

Probability Density Function Estimation by different Methods

Probability Density Function Estimation by different Methods EEE 739Q SPRIG 00 COURSE ASSIGMET REPORT Probablty Densty Functon Estmaton by dfferent Methods Vas Chandraant Rayar Abstract The am of the assgnment was to estmate the probablty densty functon (PDF of

More information

Lecture 2: Prelude to the big shrink

Lecture 2: Prelude to the big shrink Lecture 2: Prelude to the bg shrnk Last tme A slght detour wth vsualzaton tools (hey, t was the frst day... why not start out wth somethng pretty to look at?) Then, we consdered a smple 120a-style regresson

More information

Economics 130. Lecture 4 Simple Linear Regression Continued

Economics 130. Lecture 4 Simple Linear Regression Continued Economcs 130 Lecture 4 Contnued Readngs for Week 4 Text, Chapter and 3. We contnue wth addressng our second ssue + add n how we evaluate these relatonshps: Where do we get data to do ths analyss? How do

More information

MIMA Group. Chapter 2 Bayesian Decision Theory. School of Computer Science and Technology, Shandong University. Xin-Shun SDU

MIMA Group. Chapter 2 Bayesian Decision Theory. School of Computer Science and Technology, Shandong University. Xin-Shun SDU Group M D L M Chapter Bayesan Decson heory Xn-Shun Xu @ SDU School of Computer Scence and echnology, Shandong Unversty Bayesan Decson heory Bayesan decson theory s a statstcal approach to data mnng/pattern

More information

Explaining the Stein Paradox

Explaining the Stein Paradox Explanng the Sten Paradox Kwong Hu Yung 1999/06/10 Abstract Ths report offers several ratonale for the Sten paradox. Sectons 1 and defnes the multvarate normal mean estmaton problem and ntroduces Sten

More information

Hidden Markov Models & The Multivariate Gaussian (10/26/04)

Hidden Markov Models & The Multivariate Gaussian (10/26/04) CS281A/Stat241A: Statstcal Learnng Theory Hdden Markov Models & The Multvarate Gaussan (10/26/04) Lecturer: Mchael I. Jordan Scrbes: Jonathan W. Hu 1 Hdden Markov Models As a bref revew, hdden Markov models

More information

Statistical analysis using matlab. HY 439 Presented by: George Fortetsanakis

Statistical analysis using matlab. HY 439 Presented by: George Fortetsanakis Statstcal analyss usng matlab HY 439 Presented by: George Fortetsanaks Roadmap Probablty dstrbutons Statstcal estmaton Fttng data to probablty dstrbutons Contnuous dstrbutons Contnuous random varable X

More information

How its computed. y outcome data λ parameters hyperparameters. where P denotes the Laplace approximation. k i k k. Andrew B Lawson 2013

How its computed. y outcome data λ parameters hyperparameters. where P denotes the Laplace approximation. k i k k. Andrew B Lawson 2013 Andrew Lawson MUSC INLA INLA s a relatvely new tool that can be used to approxmate posteror dstrbutons n Bayesan models INLA stands for ntegrated Nested Laplace Approxmaton The approxmaton has been known

More information

See Book Chapter 11 2 nd Edition (Chapter 10 1 st Edition)

See Book Chapter 11 2 nd Edition (Chapter 10 1 st Edition) Count Data Models See Book Chapter 11 2 nd Edton (Chapter 10 1 st Edton) Count data consst of non-negatve nteger values Examples: number of drver route changes per week, the number of trp departure changes

More information

Feb 14: Spatial analysis of data fields

Feb 14: Spatial analysis of data fields Feb 4: Spatal analyss of data felds Mappng rregularly sampled data onto a regular grd Many analyss technques for geophyscal data requre the data be located at regular ntervals n space and/or tme. hs s

More information

INF 5860 Machine learning for image classification. Lecture 3 : Image classification and regression part II Anne Solberg January 31, 2018

INF 5860 Machine learning for image classification. Lecture 3 : Image classification and regression part II Anne Solberg January 31, 2018 INF 5860 Machne learnng for mage classfcaton Lecture 3 : Image classfcaton and regresson part II Anne Solberg January 3, 08 Today s topcs Multclass logstc regresson and softma Regularzaton Image classfcaton

More information

princeton univ. F 13 cos 521: Advanced Algorithm Design Lecture 3: Large deviations bounds and applications Lecturer: Sanjeev Arora

princeton univ. F 13 cos 521: Advanced Algorithm Design Lecture 3: Large deviations bounds and applications Lecturer: Sanjeev Arora prnceton unv. F 13 cos 521: Advanced Algorthm Desgn Lecture 3: Large devatons bounds and applcatons Lecturer: Sanjeev Arora Scrbe: Today s topc s devaton bounds: what s the probablty that a random varable

More information

Fourier Transform. Additive noise. Fourier Tansform. I = S + N. Noise doesn t depend on signal. We ll consider:

Fourier Transform. Additive noise. Fourier Tansform. I = S + N. Noise doesn t depend on signal. We ll consider: Flterng Announcements HW2 wll be posted later today Constructng a mosac by warpng mages. CSE252A Lecture 10a Flterng Exampel: Smoothng by Averagng Kernel: (From Bll Freeman) m=2 I Kernel sze s m+1 by m+1

More information

e i is a random error

e i is a random error Chapter - The Smple Lnear Regresson Model The lnear regresson equaton s: where + β + β e for,..., and are observable varables e s a random error How can an estmaton rule be constructed for the unknown

More information

Computation of Higher Order Moments from Two Multinomial Overdispersion Likelihood Models

Computation of Higher Order Moments from Two Multinomial Overdispersion Likelihood Models Computaton of Hgher Order Moments from Two Multnomal Overdsperson Lkelhood Models BY J. T. NEWCOMER, N. K. NEERCHAL Department of Mathematcs and Statstcs, Unversty of Maryland, Baltmore County, Baltmore,

More information

3.1 Expectation of Functions of Several Random Variables. )' be a k-dimensional discrete or continuous random vector, with joint PMF p (, E X E X1 E X

3.1 Expectation of Functions of Several Random Variables. )' be a k-dimensional discrete or continuous random vector, with joint PMF p (, E X E X1 E X Statstcs 1: Probablty Theory II 37 3 EPECTATION OF SEVERAL RANDOM VARIABLES As n Probablty Theory I, the nterest n most stuatons les not on the actual dstrbuton of a random vector, but rather on a number

More information

A Bayes Algorithm for the Multitask Pattern Recognition Problem Direct Approach

A Bayes Algorithm for the Multitask Pattern Recognition Problem Direct Approach A Bayes Algorthm for the Multtask Pattern Recognton Problem Drect Approach Edward Puchala Wroclaw Unversty of Technology, Char of Systems and Computer etworks, Wybrzeze Wyspanskego 7, 50-370 Wroclaw, Poland

More information

CS4495/6495 Introduction to Computer Vision. 3C-L3 Calibrating cameras

CS4495/6495 Introduction to Computer Vision. 3C-L3 Calibrating cameras CS4495/6495 Introducton to Computer Vson 3C-L3 Calbratng cameras Fnally (last tme): Camera parameters Projecton equaton the cumulatve effect of all parameters: M (3x4) f s x ' 1 0 0 0 c R 0 I T 3 3 3 x1

More information

OPTIMAL COMBINATION OF FOURTH ORDER STATISTICS FOR NON-CIRCULAR SOURCE SEPARATION. Christophe De Luigi and Eric Moreau

OPTIMAL COMBINATION OF FOURTH ORDER STATISTICS FOR NON-CIRCULAR SOURCE SEPARATION. Christophe De Luigi and Eric Moreau OPTIMAL COMBINATION OF FOURTH ORDER STATISTICS FOR NON-CIRCULAR SOURCE SEPARATION Chrstophe De Lug and Erc Moreau Unversty of Toulon LSEET UMR CNRS 607 av. G. Pompdou BP56 F-8362 La Valette du Var Cedex

More information

Properties of Least Squares

Properties of Least Squares Week 3 3.1 Smple Lnear Regresson Model 3. Propertes of Least Squares Estmators Y Y β 1 + β X + u weekly famly expendtures X weekly famly ncome For a gven level of x, the expected level of food expendtures

More information

Logistic Regression. CAP 5610: Machine Learning Instructor: Guo-Jun QI

Logistic Regression. CAP 5610: Machine Learning Instructor: Guo-Jun QI Logstc Regresson CAP 561: achne Learnng Instructor: Guo-Jun QI Bayes Classfer: A Generatve model odel the posteror dstrbuton P(Y X) Estmate class-condtonal dstrbuton P(X Y) for each Y Estmate pror dstrbuton

More information

Expectation Maximization Mixture Models HMMs

Expectation Maximization Mixture Models HMMs -755 Machne Learnng for Sgnal Processng Mture Models HMMs Class 9. 2 Sep 200 Learnng Dstrbutons for Data Problem: Gven a collecton of eamples from some data, estmate ts dstrbuton Basc deas of Mamum Lelhood

More information

MACHINE APPLIED MACHINE LEARNING LEARNING. Gaussian Mixture Regression

MACHINE APPLIED MACHINE LEARNING LEARNING. Gaussian Mixture Regression 11 MACHINE APPLIED MACHINE LEARNING LEARNING MACHINE LEARNING Gaussan Mture Regresson 22 MACHINE APPLIED MACHINE LEARNING LEARNING Bref summary of last week s lecture 33 MACHINE APPLIED MACHINE LEARNING

More information

Lecture 3 Stat102, Spring 2007

Lecture 3 Stat102, Spring 2007 Lecture 3 Stat0, Sprng 007 Chapter 3. 3.: Introducton to regresson analyss Lnear regresson as a descrptve technque The least-squares equatons Chapter 3.3 Samplng dstrbuton of b 0, b. Contnued n net lecture

More information

Primer on High-Order Moment Estimators

Primer on High-Order Moment Estimators Prmer on Hgh-Order Moment Estmators Ton M. Whted July 2007 The Errors-n-Varables Model We wll start wth the classcal EIV for one msmeasured regressor. The general case s n Erckson and Whted Econometrc

More information

Lecture 12: Discrete Laplacian

Lecture 12: Discrete Laplacian Lecture 12: Dscrete Laplacan Scrbe: Tanye Lu Our goal s to come up wth a dscrete verson of Laplacan operator for trangulated surfaces, so that we can use t n practce to solve related problems We are mostly

More information

Problem Set 9 Solutions

Problem Set 9 Solutions Desgn and Analyss of Algorthms May 4, 2015 Massachusetts Insttute of Technology 6.046J/18.410J Profs. Erk Demane, Srn Devadas, and Nancy Lynch Problem Set 9 Solutons Problem Set 9 Solutons Ths problem

More information

Psychology 282 Lecture #24 Outline Regression Diagnostics: Outliers

Psychology 282 Lecture #24 Outline Regression Diagnostics: Outliers Psychology 282 Lecture #24 Outlne Regresson Dagnostcs: Outlers In an earler lecture we studed the statstcal assumptons underlyng the regresson model, ncludng the followng ponts: Formal statement of assumptons.

More information

Lecture 4 Hypothesis Testing

Lecture 4 Hypothesis Testing Lecture 4 Hypothess Testng We may wsh to test pror hypotheses about the coeffcents we estmate. We can use the estmates to test whether the data rejects our hypothess. An example mght be that we wsh to

More information

Outline and Reading. Dynamic Programming. Dynamic Programming revealed. Computing Fibonacci. The General Dynamic Programming Technique

Outline and Reading. Dynamic Programming. Dynamic Programming revealed. Computing Fibonacci. The General Dynamic Programming Technique Outlne and Readng Dynamc Programmng The General Technque ( 5.3.2) -1 Knapsac Problem ( 5.3.3) Matrx Chan-Product ( 5.3.1) Dynamc Programmng verson 1.4 1 Dynamc Programmng verson 1.4 2 Dynamc Programmng

More information

BIO Lab 2: TWO-LEVEL NORMAL MODELS with school children popularity data

BIO Lab 2: TWO-LEVEL NORMAL MODELS with school children popularity data Lab : TWO-LEVEL NORMAL MODELS wth school chldren popularty data Purpose: Introduce basc two-level models for normally dstrbuted responses usng STATA. In partcular, we dscuss Random ntercept models wthout

More information

Learning with Maximum Likelihood

Learning with Maximum Likelihood Learnng wth Mamum Lelhood Note to other teachers and users of these sldes. Andrew would be delghted f you found ths source materal useful n gvng your own lectures. Feel free to use these sldes verbatm,

More information

Gaussian Mixture Models

Gaussian Mixture Models Lab Gaussan Mxture Models Lab Objectve: Understand the formulaton of Gaussan Mxture Models (GMMs) and how to estmate GMM parameters. You ve already seen GMMs as the observaton dstrbuton n certan contnuous

More information

Dr. Shalabh Department of Mathematics and Statistics Indian Institute of Technology Kanpur

Dr. Shalabh Department of Mathematics and Statistics Indian Institute of Technology Kanpur Analyss of Varance and Desgn of Experment-I MODULE VII LECTURE - 3 ANALYSIS OF COVARIANCE Dr Shalabh Department of Mathematcs and Statstcs Indan Insttute of Technology Kanpur Any scentfc experment s performed

More information

Chapter 9: Statistical Inference and the Relationship between Two Variables

Chapter 9: Statistical Inference and the Relationship between Two Variables Chapter 9: Statstcal Inference and the Relatonshp between Two Varables Key Words The Regresson Model The Sample Regresson Equaton The Pearson Correlaton Coeffcent Learnng Outcomes After studyng ths chapter,

More information

A Robust Method for Calculating the Correlation Coefficient

A Robust Method for Calculating the Correlation Coefficient A Robust Method for Calculatng the Correlaton Coeffcent E.B. Nven and C. V. Deutsch Relatonshps between prmary and secondary data are frequently quantfed usng the correlaton coeffcent; however, the tradtonal

More information