Statistical Estimation: Least Squares, Maximum Likelihood and Maximum A Posteriori Estimators
|
|
- May Rice
- 6 years ago
- Views:
Transcription
1 CS5540: Computatonal Technques for Analyzng Clncal Data Lecture 7: Statstcal Estmaton: Least Squares, Maxmum Lkelhood and Maxmum A Posteror Estmators Ashsh Raj, PhD Image Data Evaluaton and Analytcs Laboratory (IDEAL) Department of Radology Well Cornell Medcal College New York
2 Outlne Part I: Recap of Wavelet Transforms Part II : Least Squares Estmaton Part III: Maxmum Lkelhood Estmaton Part IV: Maxmum A Posteror Estmaton : Next week Note: you wll not be tested on specfc examples shown here, only on general prncples IDEA Lab, Radology, Cornell 2
3 Bass functons n WT Haar Mexcan Hat Daubeches (orthonormal) IDEA Lab, Radology, Cornell 3
4 WT n mages Images are pecewse smooth or pecewse constant Statonarty s even rarer than n 1D sgnals FT even less useful (nnd WT more attractve) 2D wavelet transforms are smple extensons of 1D WT, generally performng 1D WT along rows, then columns etc Sometmes we use 2D wavelets drectly, e.g. orthonormal Daubeches 2D wavelet IDEA Lab, Radology, Cornell 4
5 WT on mages 2D generalzaton of scaletme decomposton scale Scale 0 Scale 1 H tme Scale 2 Successve applcaton of dot product wth wavelet of ncreasng wdth. Forms a natural pyramd structure. At each scale: H = dot product of mage rows wth wavelet V = dot product of mage columns wth wavelet H-V = dot product of mage rows then columns wth wavelet IDEA Lab, Radology, Cornell V H-V 5
6 Wavelet Applcatons Many, many applcatons! Audo, mage and vdeo compresson New JPEG standard ncludes wavelet compresson FBI s fngerprnts database saved as waveletcompressed Sgnal denosng, nterpolaton, mage zoomng, texture analyss, tme-scale feature extracton In our context, WT wll be used prmarly as a feature extracton tool Remember, WT s just a change of bass, n order to extract useful nformaton whch mght otherwse not be easly seen IDEA Lab, Radology, Cornell 6
7 WT n MATLAB MATLAB has an extensve wavelet toolbox Type help wavelet n MATLAB command wndow Look at ther wavelet demo Play wth Haar, Mexcan hat and Daubeches wavelets IDEA Lab, Radology, Cornell 7
8 Project Ideas Idea 1: use WT to extract features from ECG data use these features for classfcaton Idea 2: use 2D WT to extract spato-temporal features from 3D+tme MRI data to detect tumors / classfy bengn vs malgnant tumors Idea 3: use 2D WT to denose a gven mage IDEA Lab, Radology, Cornell 8
9 Idea 3: Voxel labelng from contrast-enhanced MRI Can segment accordng to tme profle of 3D+tme contrast enhanced MR data of lver / mammography Typcal plot of tme-resolved MR sgnal of varous tssue classes Temporal models used to extract features Instead of such a smple temporal model, wavelet decomposton could provde spato-temporal features that you can use for clusterng IDEA Lab, Radology, Cornell
10 Lver tumour quantfcaton from DCE-MRI IDEA Lab, Radology, Cornell
11 Further Readng on Wavelets IDEA Lab, Radology, Cornell 11
12 Part II : Least Squares Estmaton and Examples
13 A smple Least Squares problem Lne fttng Goal: To fnd the best-ft lne representng a bunch of ponts Here: y are observatons at locaton x, Intercept and slope of lne are the unknown model parameters to be estmated Whch model parameters best ft the observed ponts? E( y nt Best ( y, m) nt =, m) ( y h( x where, m) Ths can be wrtten n matrx notaton, as θ LS = arg mn y Hθ 2 )) 2 = arg mn E( y, nt h( x) = y nt + mx What are H and θ? IDEA Lab, Radology, Cornell 13
14 Least Squares Estmator Gven lnear process y = H θ + n Least Squares estmator: θ LS = argmn y Hθ 2 Natural estmator want soluton to match observaton Does not use any nformaton about n There s a smple soluton (a.k.a. pseudo-nverse): θ LS = (H T H) -1 H T y In MATLAB, type pnv(y) IDEA Lab, Radology, Cornell 14
15 Example - estmatng T 2 decay constant n repeated spn echo MR data IDEA Lab, Radology, Cornell 15
16 Example estmatng T 2 n repeated spn echo data s(t) = s 0 e -t/t 2 Need only 2 data ponts to estmate T 2 : T 2est = [T E2 T E1 ] / ln[s(t E1 )/s(t E2 ) ] However, not good due to nose, tmng ssues In practce we have many data samples from varous echoes IDEA Lab, Radology, Cornell 16
17 Example estmatng T 2 y ln(s(t 1 )) ln(s(t 2 )) M ln(s(t n )) = 1 -t 1 1 -t 2 M 1 -t n a r H θ Least Squares estmate: θ LS = (H T H) -1 H T y T 2 = 1/r LS IDEA Lab, Radology, Cornell 17
18 Estmaton example - Denosng Suppose we have a nosy MR mage y, and wsh to obtan the noseless mage x, where y = x + n Can we use LSE to fnd x? Try: H = I, θ = x n the lnear model LS estmator smply gves x = y! we need a more powerful model Suppose the mage x can be approxmated by a polynomal,.e. a mxture of 1 st p powers of r: x = Σ =0 p a r IDEA Lab, Radology, Cornell 18
19 y Example denosng y 1 y 2 M y n = 1 r 1 1 L r 1 p 1 r 2 1 L r 2 p M 1 r n 1 L r n p a 0 a 1 M a p H + n 1 n 2 M n n Least Squares estmate: θ LS = (H T H) -1 H T y θ x = Σ =0 p a r IDEA Lab, Radology, Cornell 19
20 Part III : Maxmum Lkelhood Estmaton and Examples
21 Estmaton Theory Consder a lnear process y = H θ + n y = observed data θ = set of model parameters n = addtve nose Then Estmaton s the problem of fndng the statstcally optmal θ, gven y, H and knowledge of nose propertes Medcne s full of estmaton problems IDEA Lab, Radology, Cornell 21
22 Dfferent approaches to estmaton Mnmum varance unbased estmators Least Squares Maxmum-lkelhood Maxmum entropy Maxmum a posteror uses pror nformaton about θ has no statstcal bass uses knowledge of nose PDF IDEA Lab, Radology, Cornell 22
23 Probablty vs. Statstcs Probablty: Mathematcal models of uncertanty predct outcomes Ths s the heart of probablty Models, and ther consequences What s the probablty of a model generatng some partcular data as an outcome? Statstcs: Gven an outcome, analyze dfferent models Dd ths model generate the data? From among dfferent models (or parameters), whch one generated the data? IDEA Lab, Radology, Cornell 23
24 Maxmul Lkelhood Estmator for Lne Fttng Problem What f we know somethng about the nose?.e. Pr(n) If nose not unform across samples, LS mght be ncorrect nose σ = 10 E( y nt Best ( y y h( x, m) = ( σ nt, m) ) ) 2 = arg mn E( y, nt where, m) h( x) = y nt + mx Ths can be wrtten n matrx notaton, as θ ML = arg mn W(y Hθ) 2 nose σ = 0.1 ML estmate What s W? IDEA Lab, Radology, Cornell 24
25 Defnton of lkelhood Lkelhood s a probablty model of the uncertanty n output gven a known nput The lkelhood of a hypothess s the probablty that t would have resulted n the data you saw Thnk of the data as fxed, and try to chose among the possble PDF s Often, a parameterzed famly of PDF s ML parameter estmaton IDEA Lab, Radology, Cornell 25
26 Gaussan Nose Models In lnear model we dscussed, lkelhood comes from nose statstcs Smple dea: want to ncorporate knowledge of nose statstcs If unform whte Gaussan nose: 2 n 1 n = 1 Pr( n) exp = exp 2 2 Z 2σ Z 2σ If non-unform whte Gaussan nose: Pr( n) = 1 exp Z n 2σ IDEA Lab, Radology, Cornell 26
27 Maxmum Lkelhood Estmator - Theory n = y-hθ, Pr(n) = exp(- n 2 /2σ 2 ) Therefore Pr(y for known θ) = Pr(n) Smple dea: want to maxmze Pr(y θ) - called the lkelhood functon Example 1: show that for unform ndependent Gaussan nose θ ML = arg mn y-hθ 2 Example 2: For non-unform Gaussan nose θ ML = arg mn W(y-Hθ) 2 IDEA Lab, Radology, Cornell 27
28 MLE Bottomlne: Use nose propertes to wrte Pr(y θ) Whchever θ maxmze above, s the MLE IDEA Lab, Radology, Cornell 28
29 Example Estmatng man frequency of ECG sgnal Model: y(t ) = a sn(f t ) + n What s the MLE of a, f? Pr(y θ ) = exp(-σ (y(t ) - a sn(f t ) ) 2 / 2 σ 2 ) IDEA Lab, Radology, Cornell 29
30 Maxmum Lkelhood Detecton ML s qute a general, powerful dea Same deas can be used for classfcaton and detecton of features hdden n data Example 1: Decdng whether a voxel s artery or ven There are 3 hypotheses at each voxel: Voxel s artery, or voxel s ven, or voxel s parenchyma IDEA Lab, Radology, Cornell 30
31 Example: MRA segmentaton artery/ven may have smlar ntensty at gven tme pont but dfferent tme profles wsh to segment accordng to tme profle, not sngle ntensty IDEA Lab, Radology, Cornell 31
32 Expected Result IDEA Lab, Radology, Cornell 32
33 Example: MRA segmentaton Frst: need a tme model of all segments h( t θa ) h ( v t θ ) Lets use ML prncples to see whch voxel belongs to whch model Artery: Ven: Parench: y = h( t θa ) + n y = h( t θv ) + n y = h( t θ ) + n p IDEA Lab, Radology, Cornell
34 Maxmum Lkelhood Classfcaton Artery: Ven: Paren: y = h( t θ ) + n y = h( t θ ) + n y = h( t θ ) + n a v p Pr( y θ ) IDEA Lab, Radology, Cornell Pr( y Pr( y a θ ) v θ ) p ( y exp ( y exp ( y exp = 2 = 2 h( t θa )) 2σ h( t θv )) 2σ h( t θ )) = 2 So at each voxel, the best model s one that maxmzes: 2 ( y h( t Pr( y = θ )) θ ) Pr( y θ ) = exp 2 2σ Or equvalently, mnmzes: ( y h( t θ )) 2 2σ p 2 2 2
35 Lver tumour quantfcaton from Dynamc Contrast Enhanced MRI Data: Tumor model Rabbt DCE-MR data Paramegnetc contrast agent, pathology gold standard Extract temporal features from DCE-MRI Use these features for accurate detecton and quantfcaton of tumour IDEA Lab, Radology, Cornell
36 Lver Tumour Temporal models Temporal models used to extract features Typcal plot of tme-resolved MR sgnal of varous tssue classes h( t θ ) IDEA Lab, Radology, Cornell 36
37 Lver tumour quantfcaton from DCE-MRI IDEA Lab, Radology, Cornell
38 ML Tutorals Sldes by Andrew Moore (CMU): avalable on course webpage Paper by Jae Myung, Tutoral on Maxmum Lkelhood : avalable on course webpage Max Entropy Tutoral IDEA Lab, Radology, Cornell 38
39 Next lecture (Frday) Non-parametrc densty estmaton Hstograms + varous fttng methods Nearest neghbor Parzen estmaton Next lecture (Wednesday) Maxmum A Posteror Estmators Several examples IDEA Lab, Radology, Cornell 39
40 CS5540: Computatonal Technques for Analyzng Clncal Data Lecture 7: Statstcal Estmaton: Least Squares, Maxmum Lkelhood and Maxmum A Posteror Estmators Ashsh Raj, PhD Image Data Evaluaton and Analytcs Laboratory (IDEAL) Department of Radology Well Cornell Medcal College New York
41 Falure modes of ML Lkelhood sn t the only crteron for selectng a model or parameter Though t s obvously an mportant one Bzarre models may have hgh lkelhood Consder a speedometer readng 55 MPH Lkelhood of true speed = 55 : 10% Lkelhood of speedometer stuck : 100% ML lkes fary tales In practce, exclude such hypotheses There must be a prncpled soluton IDEA Lab, Radology, Cornell 41
42 Maxmum a Posteror Estmate Ths s an example of usng an mage pror Prors are generally expressed n the form of a PDF Pr(x) Once the lkelhood L(x) and pror are known, we have complete statstcal knowledge LS/ML are suboptmal n presence of pror MAP (aka Bayesan) estmates are optmal lkelhood Bayes Theorem: posteror Pr(x y) = Pr(y x). Pr(x) pror Pr(y) IDEA Lab, Radology, Cornell 42
43 Maxmum a Posteror (Bayesan) Estmate Consder the class of lnear systems y = Hx + n Bayesan methods maxmze the posteror probablty: Pr(x y) Pr(y x). Pr(x) Pr(y x) (lkelhood functon) = exp(- y-hx 2 ) Pr(x) (pror PDF) = exp(-g(x)) Non-Bayesan: maxmze only lkelhood x est = arg mn y-hx 2 Bayesan: x est = arg mn y-hx 2 + G(x), where G(x) s obtaned from the pror dstrbuton of x If G(x) = Gx 2 Tkhonov Regularzaton IDEA Lab, Radology, Cornell 43
44 Other example of Estmaton n MR Image denosng: H = I Image deblurrng: H = convoluton mtx n mg-space Super-resoluton: H = dagonal mtx n k-space Metabolte quantfcaton n MRSI IDEA Lab, Radology, Cornell 44
45 What Is the Rght Imagng Model? y = H x + n, n s Gaussan (1) MAP Sense y = H x + n, n, x are Gaussan (2) MAP Sense = Bayesan (MAP) estmate of (2) IDEA Lab, Radology, Cornell 45
46 Intro to Bayesan Estmaton Bayesan methods maxmze the posteror probablty: Pr(x y) Pr(y x). Pr(x) Pr(y x) (lkelhood functon) = exp(- y-hx 2 ) Pr(x) (pror PDF) = exp(-g(x)) Gaussan pror: Pr(x) = exp{- ½ x T R x -1 x} MAP estmate: x est = arg mn y-hx 2 + G(x) MAP estmate for Gaussan everythng s known as Wener estmate IDEA Lab, Radology, Cornell 46
47 Regularzaton = Bayesan Estmaton! For any regularzaton scheme, ts almost always possble to formulate the correspondng MAP problem MAP = superset of regularzaton Pror model MAP Regularzaton scheme So why deal wth regularzaton?? IDEA Lab, Radology, Cornell 47
48 Lets talk about Pror Models Temporal prors: smooth tme-trajectory Sparse prors: L0, L1, L2 (=Tkhonov) Spatal Prors: most powerful for mages I recommend robust spatal prors usng Markov Felds Want prors to be general, not too specfc Ie, weak rather than strong prors IDEA Lab, Radology, Cornell 48
49 How to do regularzaton Frst model physcal property of mage, then create a pror whch captures t, then formulate MAP estmator, Then fnd a good algorthm to solve t! How NOT to do regularzaton DON T use regularzaton scheme wthout bearng on physcal property of mage! Example: L1 or L0 pror n k-space! Specfcally: deblurrng n k-space (handy b/c convoluton becomes multply) BUT: hard to mpose smoothness prors n k-space no meanng IDEA Lab, Radology, Cornell 49
50 MAP for Lne fttng problem If model estmated by ML and Pror nfo do not agree MAP s a compromse between the two LS estmate MAP estmate Most probable pror model IDEA Lab, Radology, Cornell 50
51 Mult-varate FLASH Acqure 6-10 accelerated FLASH data sets at dfferent flp angles or TR s Generate T 1 maps by fttng to: S ( ) ( * ) 1 exp TR T = exp TE T snα 2 1 cos exp α Not enough nfo n a sngle voxel Nose causes ncorrect estmates Error n flp angle vares spatally! 1 ( TR T ) 1 IDEA Lab, Radology, Cornell 51
52 Spatally Coherent T 1, ρ estmaton Frst, stack parameters from all voxels n one bg vector x Stack all observed flp angle mages n y Then we can wrte y = M (x) + n Recall M s the (nonlnear) functonal obtaned from S ( 1) ( TR T ) ( * 1 exp TR T = exp TE T ) 2 snα 1 cos α exp Solve for x by non-lnear least square fttng, PLUS spatal pror: x est = arg mn x y - M (x) 2 + µ 2 Dx 2 E(x) Makes M(x) close to y Makes x smooth Mnmze va MATLAB s lsqnonln functon How? Construct δ = [y - M (x); µ Dx]. Then E(x) = δ 2 1 IDEA Lab, Radology, Cornell 52
53 Mult-Flp Results combned ρ, T 1 n pseudocolour IDEA Lab, Radology, Cornell 53
54 Mult-Flp Results combned ρ, T 1 n pseudocolour IDEA Lab, Radology, Cornell 54
55 Spatal Prors For Images - Example Frames are tghtly dstrbuted around mean After subtractng mean, mages are close to Gaussan varance mean tme frame N f frame 1 frame 2 mean µ x (,j) envelope a(,j) Pror: -mean s µ x -local std.dev. vares as a(,j) IDEA Lab, Radology, Cornell 55
56 Spatal Prors for MR mages statonary Stochastc MR mage model: process x(,j) = µ x (,j) + a(,j). (h ** p)(,j) (1) ** denotes 2D convoluton µ x (,j) s mean mage for class p(,j) s a unt varance..d. stochastc process a(,j) s an envelope functon h(,j) smulates correlaton propertes of mage x r(τ 1, τ 2 ) = (h ** h)(τ 1, τ 2 ) x = ACp + µ (2) where A = dag(a), and C s the Toepltz matrx generated by h Can model many mportant statonary and non-statonary cases IDEA Lab, Radology, Cornell 56
57 MAP estmate for Imagng Model (3) The Wener estmate x MAP - µ x = HR x (HR x H H + R n ) -1 (y- µ y ) (3) R x, R n = covarance matrces of x and n Statonarty R x has Toepltz structure fast processng x MAP - µ x = HACC H A H ( HACC H A H H H + σ n 2 I ) -1 (y- µ y ) (4) Drect nverson prohbtve; so use CG teratve method (4) better than (3) snce A and C are O(N log N) operatons, enablng much faster processng IDEA Lab, Radology, Cornell 57
58 How to obtan estmates of A, C? Need a tranng set of full-resoluton mages x k, k = 1,,K Parallel magng doesnt provde un-alased full-res mages Approaches: 1. Use prevous full-res scans - tme consumng, need frequent updatng 2. Use SENSE-reconstructed mages for tranng set - very bad nose amplfcaton ssues for hgh speedu ps 3. Drectly estmate parameters from avalable parallel data - Alasng may cause naccuraces IDEA Lab, Radology, Cornell 58
59 MAP for Dynamc Parallel Imagng N f mages avalable for parameter estmaton! All mages well-regstered Tghtly dstrbuted around pxel mean Parameters can be estmated from alased data tme frame N f frame 2 frame 1 k y frame 1 frame 2 frame N f N f /R full-res mages k x IDEA Lab, Radology, Cornell 59
60 MAP-SENSE Prelmnary Results Scans acceleraty 5x The angogram was computed by: avg(post-contrast) avg(pre-contrast) Unaccelerated 5x faster: SENSE 5x faster: MAP-SENSE IDEA Lab, Radology, Cornell 60
61 References Smon Kay. Statstcal Sgnal Processng. Part I: Estmaton Theory. Prentce Hall 2002 Smon Kay. Statstcal Sgnal Processng. Part II: Detecton Theory. Prentce Hall 2002 Haacke et al. Fundamentals of MRI. Zh-Pe Lang and Paul Lauterbur. Prncples of MRI A Sgnal Processng Perspectve. Info on part IV: Ashsh Raj. Improvements n MRI Usng Informaton Redundancy. PhD thess, Cornell Unversty, May Webste: IDEA Lab, Radology, Cornell 61
62 Maxmum Lkelhood Estmator But f nose s jontly Gaussan wth cov. matrx C Recall C, E(nn T ). Then Pr(n) = e -½ nt C -1 n L(y θ) = ½ (y-hθ) T C -1 (y-hθ) θ ML = argmn ½ (y-hθ) T C -1 (y-hθ) Ths also has a closed form soluton θ ML = (H T C -1 H) -1 H T C -1 y If n s not Gaussan at all, ML estmators become complcated and non-lnear Fortunately, n MR nose s usually Gaussan IDEA Lab, Radology, Cornell 62
2E Pattern Recognition Solutions to Introduction to Pattern Recognition, Chapter 2: Bayesian pattern classification
E395 - Pattern Recognton Solutons to Introducton to Pattern Recognton, Chapter : Bayesan pattern classfcaton Preface Ths document s a soluton manual for selected exercses from Introducton to Pattern Recognton
More informationLecture Notes on Linear Regression
Lecture Notes on Lnear Regresson Feng L fl@sdueducn Shandong Unversty, Chna Lnear Regresson Problem In regresson problem, we am at predct a contnuous target value gven an nput feature vector We assume
More informationComposite Hypotheses testing
Composte ypotheses testng In many hypothess testng problems there are many possble dstrbutons that can occur under each of the hypotheses. The output of the source s a set of parameters (ponts n a parameter
More informationMaximum Likelihood Estimation (MLE)
Maxmum Lkelhood Estmaton (MLE) Ken Kreutz-Delgado (Nuno Vasconcelos) ECE 175A Wnter 01 UCSD Statstcal Learnng Goal: Gven a relatonshp between a feature vector x and a vector y, and d data samples (x,y
More informationP R. Lecture 4. Theory and Applications of Pattern Recognition. Dept. of Electrical and Computer Engineering /
Theory and Applcatons of Pattern Recognton 003, Rob Polkar, Rowan Unversty, Glassboro, NJ Lecture 4 Bayes Classfcaton Rule Dept. of Electrcal and Computer Engneerng 0909.40.0 / 0909.504.04 Theory & Applcatons
More informationMLE and Bayesian Estimation. Jie Tang Department of Computer Science & Technology Tsinghua University 2012
MLE and Bayesan Estmaton Je Tang Department of Computer Scence & Technology Tsnghua Unversty 01 1 Lnear Regresson? As the frst step, we need to decde how we re gong to represent the functon f. One example:
More informationENG 8801/ Special Topics in Computer Engineering: Pattern Recognition. Memorial University of Newfoundland Pattern Recognition
EG 880/988 - Specal opcs n Computer Engneerng: Pattern Recognton Memoral Unversty of ewfoundland Pattern Recognton Lecture 7 May 3, 006 http://wwwengrmunca/~charlesr Offce Hours: uesdays hursdays 8:30-9:30
More informationLecture 10 Support Vector Machines II
Lecture 10 Support Vector Machnes II 22 February 2016 Taylor B. Arnold Yale Statstcs STAT 365/665 1/28 Notes: Problem 3 s posted and due ths upcomng Frday There was an early bug n the fake-test data; fxed
More informationCS 2750 Machine Learning. Lecture 5. Density estimation. CS 2750 Machine Learning. Announcements
CS 750 Machne Learnng Lecture 5 Densty estmaton Mlos Hauskrecht mlos@cs.ptt.edu 539 Sennott Square CS 750 Machne Learnng Announcements Homework Due on Wednesday before the class Reports: hand n before
More information8/25/17. Data Modeling. Data Modeling. Data Modeling. Patrice Koehl Department of Biological Sciences National University of Singapore
8/5/17 Data Modelng Patrce Koehl Department of Bologcal Scences atonal Unversty of Sngapore http://www.cs.ucdavs.edu/~koehl/teachng/bl59 koehl@cs.ucdavs.edu Data Modelng Ø Data Modelng: least squares Ø
More informationLecture 12: Classification
Lecture : Classfcaton g Dscrmnant functons g The optmal Bayes classfer g Quadratc classfers g Eucldean and Mahalanobs metrcs g K Nearest Neghbor Classfers Intellgent Sensor Systems Rcardo Guterrez-Osuna
More informationMotion Perception Under Uncertainty. Hongjing Lu Department of Psychology University of Hong Kong
Moton Percepton Under Uncertanty Hongjng Lu Department of Psychology Unversty of Hong Kong Outlne Uncertanty n moton stmulus Correspondence problem Qualtatve fttng usng deal observer models Based on sgnal
More informationLinear Approximation with Regularization and Moving Least Squares
Lnear Approxmaton wth Regularzaton and Movng Least Squares Igor Grešovn May 007 Revson 4.6 (Revson : March 004). 5 4 3 0.5 3 3.5 4 Contents: Lnear Fttng...4. Weghted Least Squares n Functon Approxmaton...
More informationUncertainty as the Overlap of Alternate Conditional Distributions
Uncertanty as the Overlap of Alternate Condtonal Dstrbutons Olena Babak and Clayton V. Deutsch Centre for Computatonal Geostatstcs Department of Cvl & Envronmental Engneerng Unversty of Alberta An mportant
More informationGeneralized Linear Methods
Generalzed Lnear Methods 1 Introducton In the Ensemble Methods the general dea s that usng a combnaton of several weak learner one could make a better learner. More formally, assume that we have a set
More informationClustering & Unsupervised Learning
Clusterng & Unsupervsed Learnng Ken Kreutz-Delgado (Nuno Vasconcelos) ECE 175A Wnter 2012 UCSD Statstcal Learnng Goal: Gven a relatonshp between a feature vector x and a vector y, and d data samples (x,y
More informationThe conjugate prior to a Bernoulli is. A) Bernoulli B) Gaussian C) Beta D) none of the above
The conjugate pror to a Bernoull s A) Bernoull B) Gaussan C) Beta D) none of the above The conjugate pror to a Gaussan s A) Bernoull B) Gaussan C) Beta D) none of the above MAP estmates A) argmax θ p(θ
More informationTutorial 2. COMP4134 Biometrics Authentication. February 9, Jun Xu, Teaching Asistant
Tutoral 2 COMP434 ometrcs uthentcaton Jun Xu, Teachng sstant csjunxu@comp.polyu.edu.hk February 9, 207 Table of Contents Problems Problem : nswer the questons Problem 2: Power law functon Problem 3: Convoluton
More informationGlobal Sensitivity. Tuesday 20 th February, 2018
Global Senstvty Tuesday 2 th February, 28 ) Local Senstvty Most senstvty analyses [] are based on local estmates of senstvty, typcally by expandng the response n a Taylor seres about some specfc values
More informationTracking with Kalman Filter
Trackng wth Kalman Flter Scott T. Acton Vrgna Image and Vdeo Analyss (VIVA), Charles L. Brown Department of Electrcal and Computer Engneerng Department of Bomedcal Engneerng Unversty of Vrgna, Charlottesvlle,
More informationStatistical pattern recognition
Statstcal pattern recognton Bayes theorem Problem: decdng f a patent has a partcular condton based on a partcular test However, the test s mperfect Someone wth the condton may go undetected (false negatve
More informationConjugacy and the Exponential Family
CS281B/Stat241B: Advanced Topcs n Learnng & Decson Makng Conjugacy and the Exponental Famly Lecturer: Mchael I. Jordan Scrbes: Bran Mlch 1 Conjugacy In the prevous lecture, we saw conjugate prors for the
More informationLinear Feature Engineering 11
Lnear Feature Engneerng 11 2 Least-Squares 2.1 Smple least-squares Consder the followng dataset. We have a bunch of nputs x and correspondng outputs y. The partcular values n ths dataset are x y 0.23 0.19
More informationLossy Compression. Compromise accuracy of reconstruction for increased compression.
Lossy Compresson Compromse accuracy of reconstructon for ncreased compresson. The reconstructon s usually vsbly ndstngushable from the orgnal mage. Typcally, one can get up to 0:1 compresson wth almost
More informationHomework Assignment 3 Due in class, Thursday October 15
Homework Assgnment 3 Due n class, Thursday October 15 SDS 383C Statstcal Modelng I 1 Rdge regresson and Lasso 1. Get the Prostrate cancer data from http://statweb.stanford.edu/~tbs/elemstatlearn/ datasets/prostate.data.
More informationSupport Vector Machines. Vibhav Gogate The University of Texas at dallas
Support Vector Machnes Vbhav Gogate he Unversty of exas at dallas What We have Learned So Far? 1. Decson rees. Naïve Bayes 3. Lnear Regresson 4. Logstc Regresson 5. Perceptron 6. Neural networks 7. K-Nearest
More informationLectures - Week 4 Matrix norms, Conditioning, Vector Spaces, Linear Independence, Spanning sets and Basis, Null space and Range of a Matrix
Lectures - Week 4 Matrx norms, Condtonng, Vector Spaces, Lnear Independence, Spannng sets and Bass, Null space and Range of a Matrx Matrx Norms Now we turn to assocatng a number to each matrx. We could
More informationAutomatic Object Trajectory- Based Motion Recognition Using Gaussian Mixture Models
Automatc Object Trajectory- Based Moton Recognton Usng Gaussan Mxture Models Fasal I. Bashr, Ashfaq A. Khokhar, Dan Schonfeld Electrcal and Computer Engneerng, Unversty of Illnos at Chcago. Chcago, IL,
More informationChapter 11: Simple Linear Regression and Correlation
Chapter 11: Smple Lnear Regresson and Correlaton 11-1 Emprcal Models 11-2 Smple Lnear Regresson 11-3 Propertes of the Least Squares Estmators 11-4 Hypothess Test n Smple Lnear Regresson 11-4.1 Use of t-tests
More informationLecture 3: Shannon s Theorem
CSE 533: Error-Correctng Codes (Autumn 006 Lecture 3: Shannon s Theorem October 9, 006 Lecturer: Venkatesan Guruswam Scrbe: Wdad Machmouch 1 Communcaton Model The communcaton model we are usng conssts
More informationLecture 3. Ax x i a i. i i
18.409 The Behavor of Algorthms n Practce 2/14/2 Lecturer: Dan Spelman Lecture 3 Scrbe: Arvnd Sankar 1 Largest sngular value In order to bound the condton number, we need an upper bound on the largest
More informationModule 3 LOSSY IMAGE COMPRESSION SYSTEMS. Version 2 ECE IIT, Kharagpur
Module 3 LOSSY IMAGE COMPRESSION SYSTEMS Verson ECE IIT, Kharagpur Lesson 6 Theory of Quantzaton Verson ECE IIT, Kharagpur Instructonal Objectves At the end of ths lesson, the students should be able to:
More informationWeek 5: Neural Networks
Week 5: Neural Networks Instructor: Sergey Levne Neural Networks Summary In the prevous lecture, we saw how we can construct neural networks by extendng logstc regresson. Neural networks consst of multple
More informationHidden Markov Models
Hdden Markov Models Namrata Vaswan, Iowa State Unversty Aprl 24, 204 Hdden Markov Model Defntons and Examples Defntons:. A hdden Markov model (HMM) refers to a set of hdden states X 0, X,..., X t,...,
More informationLinear Classification, SVMs and Nearest Neighbors
1 CSE 473 Lecture 25 (Chapter 18) Lnear Classfcaton, SVMs and Nearest Neghbors CSE AI faculty + Chrs Bshop, Dan Klen, Stuart Russell, Andrew Moore Motvaton: Face Detecton How do we buld a classfer to dstngush
More informationEcon107 Applied Econometrics Topic 3: Classical Model (Studenmund, Chapter 4)
I. Classcal Assumptons Econ7 Appled Econometrcs Topc 3: Classcal Model (Studenmund, Chapter 4) We have defned OLS and studed some algebrac propertes of OLS. In ths topc we wll study statstcal propertes
More informationClustering & (Ken Kreutz-Delgado) UCSD
Clusterng & Unsupervsed Learnng Nuno Vasconcelos (Ken Kreutz-Delgado) UCSD Statstcal Learnng Goal: Gven a relatonshp between a feature vector x and a vector y, and d data samples (x,y ), fnd an approxmatng
More informationFinite Mixture Models and Expectation Maximization. Most slides are from: Dr. Mario Figueiredo, Dr. Anil Jain and Dr. Rong Jin
Fnte Mxture Models and Expectaton Maxmzaton Most sldes are from: Dr. Maro Fgueredo, Dr. Anl Jan and Dr. Rong Jn Recall: The Supervsed Learnng Problem Gven a set of n samples X {(x, y )},,,n Chapter 3 of
More informationThermodynamics and statistical mechanics in materials modelling II
Course MP3 Lecture 8/11/006 (JAE) Course MP3 Lecture 8/11/006 Thermodynamcs and statstcal mechancs n materals modellng II A bref résumé of the physcal concepts used n materals modellng Dr James Ellott.1
More informationCIS526: Machine Learning Lecture 3 (Sept 16, 2003) Linear Regression. Preparation help: Xiaoying Huang. x 1 θ 1 output... θ M x M
CIS56: achne Learnng Lecture 3 (Sept 6, 003) Preparaton help: Xaoyng Huang Lnear Regresson Lnear regresson can be represented by a functonal form: f(; θ) = θ 0 0 +θ + + θ = θ = 0 ote: 0 s a dummy attrbute
More informationThe Gaussian classifier. Nuno Vasconcelos ECE Department, UCSD
he Gaussan classfer Nuno Vasconcelos ECE Department, UCSD Bayesan decson theory recall that we have state of the world X observatons g decson functon L[g,y] loss of predctng y wth g Bayes decson rule s
More informationxp(x µ) = 0 p(x = 0 µ) + 1 p(x = 1 µ) = µ
CSE 455/555 Sprng 2013 Homework 7: Parametrc Technques Jason J. Corso Computer Scence and Engneerng SUY at Buffalo jcorso@buffalo.edu Solutons by Yngbo Zhou Ths assgnment does not need to be submtted and
More informationThe big picture. Outline
The bg pcture Vncent Claveau IRISA - CNRS, sldes from E. Kjak INSA Rennes Notatons classes: C = {ω = 1,.., C} tranng set S of sze m, composed of m ponts (x, ω ) per class ω representaton space: R d (=
More informationChange Detection: Current State of the Art and Future Directions
Change Detecton: Current State of the Art and Future Drectons Dapeng Olver Wu Electrcal & Computer Engneerng Unversty of Florda http://www.wu.ece.ufl.edu/ Outlne Motvaton & problem statement Change detecton
More informationChapter 2 - The Simple Linear Regression Model S =0. e i is a random error. S β2 β. This is a minimization problem. Solution is a calculus exercise.
Chapter - The Smple Lnear Regresson Model The lnear regresson equaton s: where y + = β + β e for =,..., y and are observable varables e s a random error How can an estmaton rule be constructed for the
More informationCommunication with AWGN Interference
Communcaton wth AWG Interference m {m } {p(m } Modulator s {s } r=s+n Recever ˆm AWG n m s a dscrete random varable(rv whch takes m wth probablty p(m. Modulator maps each m nto a waveform sgnal s m=m
More informationMaximum Likelihood Estimation of Binary Dependent Variables Models: Probit and Logit. 1. General Formulation of Binary Dependent Variables Models
ECO 452 -- OE 4: Probt and Logt Models ECO 452 -- OE 4 Maxmum Lkelhood Estmaton of Bnary Dependent Varables Models: Probt and Logt hs note demonstrates how to formulate bnary dependent varables models
More informationStat 543 Exam 2 Spring 2016
Stat 543 Exam 2 Sprng 206 I have nether gven nor receved unauthorzed assstance on ths exam. Name Sgned Date Name Prnted Ths Exam conssts of questons. Do at least 0 of the parts of the man exam. I wll score
More informationUsing T.O.M to Estimate Parameter of distributions that have not Single Exponential Family
IOSR Journal of Mathematcs IOSR-JM) ISSN: 2278-5728. Volume 3, Issue 3 Sep-Oct. 202), PP 44-48 www.osrjournals.org Usng T.O.M to Estmate Parameter of dstrbutons that have not Sngle Exponental Famly Jubran
More informationFor now, let us focus on a specific model of neurons. These are simplified from reality but can achieve remarkable results.
Neural Networks : Dervaton compled by Alvn Wan from Professor Jtendra Malk s lecture Ths type of computaton s called deep learnng and s the most popular method for many problems, such as computer vson
More informationSIO 224. m(r) =(ρ(r),k s (r),µ(r))
SIO 224 1. A bref look at resoluton analyss Here s some background for the Masters and Gubbns resoluton paper. Global Earth models are usually found teratvely by assumng a startng model and fndng small
More informationChat eld, C. and A.J.Collins, Introduction to multivariate analysis. Chapman & Hall, 1980
MT07: Multvarate Statstcal Methods Mke Tso: emal mke.tso@manchester.ac.uk Webpage for notes: http://www.maths.manchester.ac.uk/~mkt/new_teachng.htm. Introducton to multvarate data. Books Chat eld, C. and
More informationMATH 829: Introduction to Data Mining and Analysis The EM algorithm (part 2)
1/16 MATH 829: Introducton to Data Mnng and Analyss The EM algorthm (part 2) Domnque Gullot Departments of Mathematcal Scences Unversty of Delaware Aprl 20, 2016 Recall 2/16 We are gven ndependent observatons
More informationLimited Dependent Variables
Lmted Dependent Varables. What f the left-hand sde varable s not a contnuous thng spread from mnus nfnty to plus nfnty? That s, gven a model = f (, β, ε, where a. s bounded below at zero, such as wages
More informationMarkov Chain Monte Carlo (MCMC), Gibbs Sampling, Metropolis Algorithms, and Simulated Annealing Bioinformatics Course Supplement
Markov Chan Monte Carlo MCMC, Gbbs Samplng, Metropols Algorthms, and Smulated Annealng 2001 Bonformatcs Course Supplement SNU Bontellgence Lab http://bsnuackr/ Outlne! Markov Chan Monte Carlo MCMC! Metropols-Hastngs
More informationSpace of ML Problems. CSE 473: Artificial Intelligence. Parameter Estimation and Bayesian Networks. Learning Topics
/7/7 CSE 73: Artfcal Intellgence Bayesan - Learnng Deter Fox Sldes adapted from Dan Weld, Jack Breese, Dan Klen, Daphne Koller, Stuart Russell, Andrew Moore & Luke Zettlemoyer What s Beng Learned? Space
More informationMachine learning: Density estimation
CS 70 Foundatons of AI Lecture 3 Machne learnng: ensty estmaton Mlos Hauskrecht mlos@cs.ptt.edu 539 Sennott Square ata: ensty estmaton {.. n} x a vector of attrbute values Objectve: estmate the model of
More informationStat 543 Exam 2 Spring 2016
Stat 543 Exam 2 Sprng 2016 I have nether gven nor receved unauthorzed assstance on ths exam. Name Sgned Date Name Prnted Ths Exam conssts of 11 questons. Do at least 10 of the 11 parts of the man exam.
More informationsince [1-( 0+ 1x1i+ 2x2 i)] [ 0+ 1x1i+ assumed to be a reasonable approximation
Econ 388 R. Butler 204 revsons Lecture 4 Dummy Dependent Varables I. Lnear Probablty Model: the Regresson model wth a dummy varables as the dependent varable assumpton, mplcaton regular multple regresson
More informationDr. Shalabh Department of Mathematics and Statistics Indian Institute of Technology Kanpur
Analyss of Varance and Desgn of Exerments-I MODULE III LECTURE - 2 EXPERIMENTAL DESIGN MODELS Dr. Shalabh Deartment of Mathematcs and Statstcs Indan Insttute of Technology Kanur 2 We consder the models
More informationComparison of Regression Lines
STATGRAPHICS Rev. 9/13/2013 Comparson of Regresson Lnes Summary... 1 Data Input... 3 Analyss Summary... 4 Plot of Ftted Model... 6 Condtonal Sums of Squares... 6 Analyss Optons... 7 Forecasts... 8 Confdence
More informationInstance-Based Learning (a.k.a. memory-based learning) Part I: Nearest Neighbor Classification
Instance-Based earnng (a.k.a. memory-based learnng) Part I: Nearest Neghbor Classfcaton Note to other teachers and users of these sldes. Andrew would be delghted f you found ths source materal useful n
More informationCSci 6974 and ECSE 6966 Math. Tech. for Vision, Graphics and Robotics Lecture 21, April 17, 2006 Estimating A Plane Homography
CSc 6974 and ECSE 6966 Math. Tech. for Vson, Graphcs and Robotcs Lecture 21, Aprl 17, 2006 Estmatng A Plane Homography Overvew We contnue wth a dscusson of the major ssues, usng estmaton of plane projectve
More informationWhy Bayesian? 3. Bayes and Normal Models. State of nature: class. Decision rule. Rev. Thomas Bayes ( ) Bayes Theorem (yes, the famous one)
Why Bayesan? 3. Bayes and Normal Models Alex M. Martnez alex@ece.osu.edu Handouts Handoutsfor forece ECE874 874Sp Sp007 If all our research (n PR was to dsappear and you could only save one theory, whch
More informationProbability Density Function Estimation by different Methods
EEE 739Q SPRIG 00 COURSE ASSIGMET REPORT Probablty Densty Functon Estmaton by dfferent Methods Vas Chandraant Rayar Abstract The am of the assgnment was to estmate the probablty densty functon (PDF of
More informationLecture 2: Prelude to the big shrink
Lecture 2: Prelude to the bg shrnk Last tme A slght detour wth vsualzaton tools (hey, t was the frst day... why not start out wth somethng pretty to look at?) Then, we consdered a smple 120a-style regresson
More informationEconomics 130. Lecture 4 Simple Linear Regression Continued
Economcs 130 Lecture 4 Contnued Readngs for Week 4 Text, Chapter and 3. We contnue wth addressng our second ssue + add n how we evaluate these relatonshps: Where do we get data to do ths analyss? How do
More informationMIMA Group. Chapter 2 Bayesian Decision Theory. School of Computer Science and Technology, Shandong University. Xin-Shun SDU
Group M D L M Chapter Bayesan Decson heory Xn-Shun Xu @ SDU School of Computer Scence and echnology, Shandong Unversty Bayesan Decson heory Bayesan decson theory s a statstcal approach to data mnng/pattern
More informationExplaining the Stein Paradox
Explanng the Sten Paradox Kwong Hu Yung 1999/06/10 Abstract Ths report offers several ratonale for the Sten paradox. Sectons 1 and defnes the multvarate normal mean estmaton problem and ntroduces Sten
More informationHidden Markov Models & The Multivariate Gaussian (10/26/04)
CS281A/Stat241A: Statstcal Learnng Theory Hdden Markov Models & The Multvarate Gaussan (10/26/04) Lecturer: Mchael I. Jordan Scrbes: Jonathan W. Hu 1 Hdden Markov Models As a bref revew, hdden Markov models
More informationStatistical analysis using matlab. HY 439 Presented by: George Fortetsanakis
Statstcal analyss usng matlab HY 439 Presented by: George Fortetsanaks Roadmap Probablty dstrbutons Statstcal estmaton Fttng data to probablty dstrbutons Contnuous dstrbutons Contnuous random varable X
More informationHow its computed. y outcome data λ parameters hyperparameters. where P denotes the Laplace approximation. k i k k. Andrew B Lawson 2013
Andrew Lawson MUSC INLA INLA s a relatvely new tool that can be used to approxmate posteror dstrbutons n Bayesan models INLA stands for ntegrated Nested Laplace Approxmaton The approxmaton has been known
More informationSee Book Chapter 11 2 nd Edition (Chapter 10 1 st Edition)
Count Data Models See Book Chapter 11 2 nd Edton (Chapter 10 1 st Edton) Count data consst of non-negatve nteger values Examples: number of drver route changes per week, the number of trp departure changes
More informationFeb 14: Spatial analysis of data fields
Feb 4: Spatal analyss of data felds Mappng rregularly sampled data onto a regular grd Many analyss technques for geophyscal data requre the data be located at regular ntervals n space and/or tme. hs s
More informationINF 5860 Machine learning for image classification. Lecture 3 : Image classification and regression part II Anne Solberg January 31, 2018
INF 5860 Machne learnng for mage classfcaton Lecture 3 : Image classfcaton and regresson part II Anne Solberg January 3, 08 Today s topcs Multclass logstc regresson and softma Regularzaton Image classfcaton
More informationprinceton univ. F 13 cos 521: Advanced Algorithm Design Lecture 3: Large deviations bounds and applications Lecturer: Sanjeev Arora
prnceton unv. F 13 cos 521: Advanced Algorthm Desgn Lecture 3: Large devatons bounds and applcatons Lecturer: Sanjeev Arora Scrbe: Today s topc s devaton bounds: what s the probablty that a random varable
More informationFourier Transform. Additive noise. Fourier Tansform. I = S + N. Noise doesn t depend on signal. We ll consider:
Flterng Announcements HW2 wll be posted later today Constructng a mosac by warpng mages. CSE252A Lecture 10a Flterng Exampel: Smoothng by Averagng Kernel: (From Bll Freeman) m=2 I Kernel sze s m+1 by m+1
More informatione i is a random error
Chapter - The Smple Lnear Regresson Model The lnear regresson equaton s: where + β + β e for,..., and are observable varables e s a random error How can an estmaton rule be constructed for the unknown
More informationComputation of Higher Order Moments from Two Multinomial Overdispersion Likelihood Models
Computaton of Hgher Order Moments from Two Multnomal Overdsperson Lkelhood Models BY J. T. NEWCOMER, N. K. NEERCHAL Department of Mathematcs and Statstcs, Unversty of Maryland, Baltmore County, Baltmore,
More information3.1 Expectation of Functions of Several Random Variables. )' be a k-dimensional discrete or continuous random vector, with joint PMF p (, E X E X1 E X
Statstcs 1: Probablty Theory II 37 3 EPECTATION OF SEVERAL RANDOM VARIABLES As n Probablty Theory I, the nterest n most stuatons les not on the actual dstrbuton of a random vector, but rather on a number
More informationA Bayes Algorithm for the Multitask Pattern Recognition Problem Direct Approach
A Bayes Algorthm for the Multtask Pattern Recognton Problem Drect Approach Edward Puchala Wroclaw Unversty of Technology, Char of Systems and Computer etworks, Wybrzeze Wyspanskego 7, 50-370 Wroclaw, Poland
More informationCS4495/6495 Introduction to Computer Vision. 3C-L3 Calibrating cameras
CS4495/6495 Introducton to Computer Vson 3C-L3 Calbratng cameras Fnally (last tme): Camera parameters Projecton equaton the cumulatve effect of all parameters: M (3x4) f s x ' 1 0 0 0 c R 0 I T 3 3 3 x1
More informationOPTIMAL COMBINATION OF FOURTH ORDER STATISTICS FOR NON-CIRCULAR SOURCE SEPARATION. Christophe De Luigi and Eric Moreau
OPTIMAL COMBINATION OF FOURTH ORDER STATISTICS FOR NON-CIRCULAR SOURCE SEPARATION Chrstophe De Lug and Erc Moreau Unversty of Toulon LSEET UMR CNRS 607 av. G. Pompdou BP56 F-8362 La Valette du Var Cedex
More informationProperties of Least Squares
Week 3 3.1 Smple Lnear Regresson Model 3. Propertes of Least Squares Estmators Y Y β 1 + β X + u weekly famly expendtures X weekly famly ncome For a gven level of x, the expected level of food expendtures
More informationLogistic Regression. CAP 5610: Machine Learning Instructor: Guo-Jun QI
Logstc Regresson CAP 561: achne Learnng Instructor: Guo-Jun QI Bayes Classfer: A Generatve model odel the posteror dstrbuton P(Y X) Estmate class-condtonal dstrbuton P(X Y) for each Y Estmate pror dstrbuton
More informationExpectation Maximization Mixture Models HMMs
-755 Machne Learnng for Sgnal Processng Mture Models HMMs Class 9. 2 Sep 200 Learnng Dstrbutons for Data Problem: Gven a collecton of eamples from some data, estmate ts dstrbuton Basc deas of Mamum Lelhood
More informationMACHINE APPLIED MACHINE LEARNING LEARNING. Gaussian Mixture Regression
11 MACHINE APPLIED MACHINE LEARNING LEARNING MACHINE LEARNING Gaussan Mture Regresson 22 MACHINE APPLIED MACHINE LEARNING LEARNING Bref summary of last week s lecture 33 MACHINE APPLIED MACHINE LEARNING
More informationLecture 3 Stat102, Spring 2007
Lecture 3 Stat0, Sprng 007 Chapter 3. 3.: Introducton to regresson analyss Lnear regresson as a descrptve technque The least-squares equatons Chapter 3.3 Samplng dstrbuton of b 0, b. Contnued n net lecture
More informationPrimer on High-Order Moment Estimators
Prmer on Hgh-Order Moment Estmators Ton M. Whted July 2007 The Errors-n-Varables Model We wll start wth the classcal EIV for one msmeasured regressor. The general case s n Erckson and Whted Econometrc
More informationLecture 12: Discrete Laplacian
Lecture 12: Dscrete Laplacan Scrbe: Tanye Lu Our goal s to come up wth a dscrete verson of Laplacan operator for trangulated surfaces, so that we can use t n practce to solve related problems We are mostly
More informationProblem Set 9 Solutions
Desgn and Analyss of Algorthms May 4, 2015 Massachusetts Insttute of Technology 6.046J/18.410J Profs. Erk Demane, Srn Devadas, and Nancy Lynch Problem Set 9 Solutons Problem Set 9 Solutons Ths problem
More informationPsychology 282 Lecture #24 Outline Regression Diagnostics: Outliers
Psychology 282 Lecture #24 Outlne Regresson Dagnostcs: Outlers In an earler lecture we studed the statstcal assumptons underlyng the regresson model, ncludng the followng ponts: Formal statement of assumptons.
More informationLecture 4 Hypothesis Testing
Lecture 4 Hypothess Testng We may wsh to test pror hypotheses about the coeffcents we estmate. We can use the estmates to test whether the data rejects our hypothess. An example mght be that we wsh to
More informationOutline and Reading. Dynamic Programming. Dynamic Programming revealed. Computing Fibonacci. The General Dynamic Programming Technique
Outlne and Readng Dynamc Programmng The General Technque ( 5.3.2) -1 Knapsac Problem ( 5.3.3) Matrx Chan-Product ( 5.3.1) Dynamc Programmng verson 1.4 1 Dynamc Programmng verson 1.4 2 Dynamc Programmng
More informationBIO Lab 2: TWO-LEVEL NORMAL MODELS with school children popularity data
Lab : TWO-LEVEL NORMAL MODELS wth school chldren popularty data Purpose: Introduce basc two-level models for normally dstrbuted responses usng STATA. In partcular, we dscuss Random ntercept models wthout
More informationLearning with Maximum Likelihood
Learnng wth Mamum Lelhood Note to other teachers and users of these sldes. Andrew would be delghted f you found ths source materal useful n gvng your own lectures. Feel free to use these sldes verbatm,
More informationGaussian Mixture Models
Lab Gaussan Mxture Models Lab Objectve: Understand the formulaton of Gaussan Mxture Models (GMMs) and how to estmate GMM parameters. You ve already seen GMMs as the observaton dstrbuton n certan contnuous
More informationDr. Shalabh Department of Mathematics and Statistics Indian Institute of Technology Kanpur
Analyss of Varance and Desgn of Experment-I MODULE VII LECTURE - 3 ANALYSIS OF COVARIANCE Dr Shalabh Department of Mathematcs and Statstcs Indan Insttute of Technology Kanpur Any scentfc experment s performed
More informationChapter 9: Statistical Inference and the Relationship between Two Variables
Chapter 9: Statstcal Inference and the Relatonshp between Two Varables Key Words The Regresson Model The Sample Regresson Equaton The Pearson Correlaton Coeffcent Learnng Outcomes After studyng ths chapter,
More informationA Robust Method for Calculating the Correlation Coefficient
A Robust Method for Calculatng the Correlaton Coeffcent E.B. Nven and C. V. Deutsch Relatonshps between prmary and secondary data are frequently quantfed usng the correlaton coeffcent; however, the tradtonal
More information