Learning with Maximum Likelihood

Size: px
Start display at page:

Download "Learning with Maximum Likelihood"

Transcription

1 Learnng wth Mamum Lelhood Note to other teachers and users of these sldes. Andrew would be delghted f you found ths source materal useful n gvng your own lectures. Feel free to use these sldes verbatm, or to modfy them to ft your own needs. PowerPont orgnals are avalable. If you mae use of a sgnfcant porton of these sldes n your own lecture, please nclude ths message, or the followng ln to the source repostory of Andrew s tutorals: Comments and correctons gratefully receved. Andrew W. Moore Professor School of Computer Scence Carnege Mellon Unversty awm@cs.cmu.edu Copyrght 00, 004, Andrew W. Moore Sep 6th, 00 Mamum Lelhood learnng of Gaussans for Data Mnng Why we should care Learnng Unvarate Gaussans Learnng Multvarate Gaussans What s a based estmator? Bayesan Learnng of Gaussans Copyrght 00, 004, Andrew W. Moore Mamum Lelhood: Slde

2 Why we should care Mamum Lelhood Estmaton s a very very very very fundamental part of data analyss. MLE for Gaussans s tranng wheels for our future technques Learnng Gaussans s more useful than you mght guess Copyrght 00, 004, Andrew W. Moore Mamum Lelhood: Slde 3 Learnng Gaussans from Data Suppose you have,, ~ (..d) N(, ) But you don t now (you do now ) MLE: For whch s,, most lely? MAP: Whch mamzes p(,,, )? Copyrght 00, 004, Andrew W. Moore Mamum Lelhood: Slde 4

3 Learnng Gaussans from Data Suppose you have,, ~(..d) N(, ) But you don t now Sneer (you do now ) MLE: For whch s,, most lely? MAP: Whch mamzes p(,,, )? Copyrght 00, 004, Andrew W. Moore Mamum Lelhood: Slde 5 Learnng Gaussans from Data Suppose you have,, ~(..d) N(, ) But you don t now Sneer (you do now ) MLE: For whch s,, most lely? MAP: Whch mamzes p(,,, )? Despte ths, we ll spend 95% of our tme on MLE. Why? Wat and see Copyrght 00, 004, Andrew W. Moore Mamum Lelhood: Slde 6 3

4 MLE for unvarate Gaussan Suppose you have,, ~(..d) N(, ) But you don t now (you do now ) MLE: For whch s,, most lely? arg ma p(,,..., ) Copyrght 00, 004, Andrew W. Moore Mamum Lelhood: Slde 7 Algebra Euphora arg ma p(,,..., ) (by..d) (monotoncty of log) (plug n formula for Gaussan) (after smplfcaton) Copyrght 00, 004, Andrew W. Moore Mamum Lelhood: Slde 8 4

5 arg ma p( Algebra Euphora arg ma p(,,..., ) arg ma arg ma log p( π, ( arg mn ( ), ) ) ) (by..d) (monotoncty of log) (plug n formula for Gaussan) (after smplfcaton) Copyrght 00, 004, Andrew W. Moore Mamum Lelhood: Slde 9 Intermsson: A General Scalar MLE strategy Tas: Fnd MLE θ assumng nown form for p(data θ,stuff). Wrte LL log P(Data θ,stuff). Wor out / θ usng hgh-school calculus 3. Set / θ0 for a mamum, creatng an equaton n terms of θ 4. Solve t* 5. Chec that you ve found a mamum rather than a mnmum or saddle-pont, and be careful f θ s constraned *Ths s a perfect eample of somethng that wors perfectly n all tetboo eamples and usually nvolves surprsng pan f you need t for somethng new. Copyrght 00, 004, Andrew W. Moore Mamum Lelhood: Slde 0 5

6 The MLE arg ma p(,,..., ) arg mn ( ) s.t. 0 (what?) Copyrght 00, 004, Andrew W. Moore Mamum Lelhood: Slde The MLE arg ma p(,,..., ) arg mn ( ) s.t. 0 Thus Copyrght 00, 004, Andrew W. Moore Mamum Lelhood: Slde ( ) ( ) 6

7 Laws-a-lawdy! The best estmate of the mean of a dstrbuton s the mean of the sample! At frst sght: Ths nd of pedantc, algebra-flled and ultmately unsurprsng fact s eactly the reason people throw down ther Statstcs boo and pc up ther Agent Based Evolutonary Data Mnng Usng The Neuro-Fuzz Transform boo. Copyrght 00, 004, Andrew W. Moore Mamum Lelhood: Slde 3 A General MLE strategy Suppose θ (θ, θ,, θ n ) T s a vector of parameters. Tas: Fnd MLE θ assumng nown form for p(data θ,stuff). Wrte LL log P(Data θ,stuff). Wor out / θ usng hgh-school calculus θ θ θ M θ n Copyrght 00, 004, Andrew W. Moore Mamum Lelhood: Slde 4 7

8 A General MLE strategy Suppose θ (θ, θ,, θ n ) T s a vector of parameters. Tas: Fnd MLE θ assumng nown form for p(data θ,stuff). Wrte LL log P(Data θ,stuff). Wor out / θ usng hgh-school calculus 3. Solve the set of smultaneous equatons θ θ M θ n Copyrght 00, 004, Andrew W. Moore Mamum Lelhood: Slde 5 A General MLE strategy Suppose θ (θ, θ,, θ n ) T s a vector of parameters. Tas: Fnd MLE θ assumng nown form for p(data θ,stuff). Wrte LL log P(Data θ,stuff). Wor out / θ usng hgh-school calculus 3. Solve the set of smultaneous equatons 0 θ 0 θ M 0 θ n 4. Chec that you re at a mamum Copyrght 00, 004, Andrew W. Moore Mamum Lelhood: Slde 6 8

9 A General MLE strategy Suppose θ (θ, θ,, θ n ) T s a vector of parameters. Tas: Fnd MLE θ assumng nown form for p(data θ,stuff). Wrte LL log P(Data θ,stuff). Wor out / θ usng hgh-school calculus 3. Solve the set of smultaneous equatons If you can t solve them, what should you do? 0 θ 0 θ M 0 θ n 4. Chec that you re at a mamum Copyrght 00, 004, Andrew W. Moore Mamum Lelhood: Slde 7 MLE for unvarate Gaussan Suppose you have,, ~(..d) N(, ) But you don t now or MLE: For whch θ (, ) s,, most lely? log p(,,..., ) (log π + log ) ( ) ( ) + ( ) 4 Copyrght 00, 004, Andrew W. Moore Mamum Lelhood: Slde 8 9

10 MLE for unvarate Gaussan Suppose you have,, ~(..d) N(, ) But you don t now or MLE: For whch θ (, ) s,, most lely? log p(,,..., ) (log π + log ) ( ) 0 ( ) 0 + ( ) 4 Copyrght 00, 004, Andrew W. Moore Mamum Lelhood: Slde 9 MLE for unvarate Gaussan Suppose you have,, ~(..d) N(, ) But you don t now or MLE: For whch θ (, ) s,, most lely? log p(,,..., ) (log π + log ) ( ) 0 ( ) 0 + ( ) 4 what? Copyrght 00, 004, Andrew W. Moore Mamum Lelhood: Slde 0 0

11 MLE for unvarate Gaussan Suppose you have,, ~(..d) N(, ) But you don t now or MLE: For whch θ (, ) s,, most lely? ( ) Copyrght 00, 004, Andrew W. Moore Mamum Lelhood: Slde Unbased Estmators An estmator of a parameter s unbased f the epected value of the estmate s the same as the true value of the parameters. If,, ~(..d) N(, ) then E[ ] E s unbased Copyrght 00, 004, Andrew W. Moore Mamum Lelhood: Slde

12 Copyrght 00, 004, Andrew W. Moore Mamum Lelhood: Slde 3 Based Estmators An estmator of a parameter s based f the epected value of the estmate s dfferent from the true value of the parameters. If,, ~(..d) N(, ) then [ ] ) ( j j E E E s based Copyrght 00, 004, Andrew W. Moore Mamum Lelhood: Slde 4 MLE Varance Bas If,, ~(..d) N(, ) then [ ] E E j j Intuton chec: consder the case of Why should our guts epect that would be an underestmate of true? How could you prove that?

13 Unbased estmate of Varance If,, ~(..d) N(, ) then E [ ] E j j So defne unbased [ ] So E unbased Copyrght 00, 004, Andrew W. Moore Mamum Lelhood: Slde 5 Unbased estmate of Varance If,, ~(..d) N(, ) then E [ ] E j j So defne unbased [ ] So E unbased unbased ( ) Copyrght 00, 004, Andrew W. Moore Mamum Lelhood: Slde 6 3

14 Unbasedtude dscusson Whch s best? ( unbased ( ) ) Answer: It depends on the tas And doesn t mae much dfference once --> large Copyrght 00, 004, Andrew W. Moore Mamum Lelhood: Slde 7 Don t get too ected about beng unbased Assume,, ~(..d) N(, ) Suppose we had these estmators for the mean suboptmal + 7 crap Are ether of these unbased? Wll ether of them asymptote to the correct value as gets large? Whch s more useful? Copyrght 00, 004, Andrew W. Moore Mamum Lelhood: Slde 8 4

15 MLE for m-dmensonal Gaussan Suppose you have,, ~(..d) N(,Σ) But you don t now or Σ MLE: For whch θ (,Σ) s,, most lely? Σ ( )( ) T Copyrght 00, 004, Andrew W. Moore Mamum Lelhood: Slde 9 MLE for m-dmensonal Gaussan Suppose you have,, ~(..d) N(,Σ) But you don t now or Σ MLE: For whch θ (,Σ) s,, most lely? Σ ( )( ) T Where m And s value of the th component of (the th attrbute of the th record) And s the th component of Copyrght 00, 004, Andrew W. Moore Mamum Lelhood: Slde 30 5

16 MLE for m-dmensonal Gaussan Suppose you have,, ~(..d) N(,Σ) But you don t now or Σ MLE: For whch θ (,Σ) s,, most lely? Σ ( )( ) j Copyrght 00, 004, Andrew W. Moore Mamum Lelhood: Slde 3 T ( )( j ) j Where m, j m And s value of the th component of (the th attrbute of the th record) And j s the (,j) th component of Σ MLE for m-dmensonal Gaussan Suppose you have,, ~(..d) N(,Σ) But you don t now or Σ MLE: For whch θ (,Σ) s Note, how, Σ most s forced lely? to be Σ ( )( ) Σ unbased Σ Copyrght 00, 004, Andrew W. Moore Mamum Lelhood: Slde 3 T Q: How would you prove ths? A: Just plug through the MLE recpe. symmetrc non-negatve defnte Note the unbased case How many dataponts would you need before the Gaussan has a chance of beng non-degenerate? ( )( ) T 6

17 Confdence ntervals We need to tal We need to dscuss how accurate we epect and Σ to be as a functon of And we need to consder how to estmate these accuraces from data Analytcally * Non-parametrcally (usng randomzaton and bootstrappng) * But we won t. Not yet. *Wll be dscussed n future Andrew lectures just before we need ths technology. Copyrght 00, 004, Andrew W. Moore Mamum Lelhood: Slde 33 Structural error Actually, we need to tal about somethng else too.. What f we do all ths analyss when the true dstrbuton s n fact not Gaussan? How can we tell? * How can we survve? * *Wll be dscussed n future Andrew lectures just before we need ths technology. Copyrght 00, 004, Andrew W. Moore Mamum Lelhood: Slde 34 7

18 Gaussan MLE n acton Usng 39 cars from the MPG UCI dataset suppled by oss Qunlan Copyrght 00, 004, Andrew W. Moore Mamum Lelhood: Slde 35 Data-starved Gaussan MLE Usng three subsets of MPG. Each subset has 6 randomly-chosen cars. Copyrght 00, 004, Andrew W. Moore Mamum Lelhood: Slde 36 8

19 Bvarate MLE n acton Copyrght 00, 004, Andrew W. Moore Mamum Lelhood: Slde 37 Multvarate MLE Covarance matrces are not ectng to loo at Copyrght 00, 004, Andrew W. Moore Mamum Lelhood: Slde 38 9

20 Beng Bayesan: MAP estmates for Gaussans Suppose you have,, ~(..d) N(,Σ) But you don t now or Σ MAP: Whch (,Σ) mamzes p(,σ,, )? Step : Put a pror on (,Σ) Copyrght 00, 004, Andrew W. Moore Mamum Lelhood: Slde 39 Beng Bayesan: MAP estmates for Gaussans Suppose you have,, ~(..d) N(,Σ) But you don t now or Σ MAP: Whch (,Σ) mamzes p(,σ,, )? Step : Put a pror on (,Σ) Step a: Put a pror on Σ (ν 0 -m-) Σ ~ IW(ν 0, (ν 0 -m-) Σ 0 ) Ths thng s called the Inverse-Wshart dstrbuton. A PDF over SPD matrces! Copyrght 00, 004, Andrew W. Moore Mamum Lelhood: Slde 40 0

21 Beng Bayesan: MAP estmates for Gaussans ν Suppose 0 small: I am not sure you have,, ~(..d) N(,Σ) about my guess of Σ 0 Σ But you don t now or Σ 0 : (oughly) my best guess of Σ MAP: ν 0 large: Whch I m pretty (,Σ) sure mamzes p(,σ,, )? about my guess of Σ 0 Ε[Σ ] Σ 0 Step : Put a pror on (,Σ) Step a: Put a pror on Σ (ν 0 -m-) Σ ~ IW(ν 0, (ν 0 -m-) Σ 0 ) Ths thng s called the Inverse-Wshart dstrbuton. A PDF over SPD matrces! Copyrght 00, 004, Andrew W. Moore Mamum Lelhood: Slde 4 Beng Bayesan: MAP estmates for Gaussans Suppose you have,, ~(..d) N(,Σ) But you don t now or Σ MAP: Whch (,Σ) mamzes p(,σ,, )? Step : Put a pror on (,Σ) Step a: Put a pror on Σ (ν 0 -m-)σ ~ IW(ν 0, (ν 0 -m-)σ 0 ) Step b: Put a pror on Σ Σ ~ N( 0, Σ / κ 0 ) Together, Σ and Σ defne a jont dstrbuton on (,Σ) Copyrght 00, 004, Andrew W. Moore Mamum Lelhood: Slde 4

22 Beng Bayesan: MAP estmates for Gaussans Suppose you have,, ~(..d) N(,Σ) But you don t now or Σ κ 0 small: I am not sure MAP: Whch (,Σ) mamzes about p(,σ my guess, of, 0 )? 0 Step : My : best Put guess a pror of on (,Σ) Step E[ a: ] Put a 0 pror on Σ (ν 0 -m-)σ ~ IW(ν 0, (ν 0 -m-)σ 0 ) Step b: Put a pror on Σ Σ ~ N( 0, Σ / κ 0 ) Notce how we are forced to epress our gnorance of proportonally to Σ κ 0 large: I m pretty sure about my guess of 0 Together, Σ and Σ defne a jont dstrbuton on (,Σ) Copyrght 00, 004, Andrew W. Moore Mamum Lelhood: Slde 43 Beng Bayesan: MAP estmates for Gaussans Suppose you have,, ~(..d) N(,Σ) But you don t now or Σ MAP: Whch (,Σ) mamzes p(,σ,, )? Step : Put a pror on (,Σ) Step a: Put a pror on Σ (ν 0 -m-)σ ~ IW(ν 0, (ν 0 -m-)σ 0 ) Step b: Put a pror on Σ Σ ~ N( 0, Σ / κ 0 ) Why do we use ths form of pror? Copyrght 00, 004, Andrew W. Moore Mamum Lelhood: Slde 44

23 Beng Bayesan: MAP estmates for Gaussans Suppose you have,, ~(..d) N(,Σ) But you don t now or Σ MAP: Whch (,Σ) mamzes p(,σ,, )? Step : Put a pror on (,Σ) Step a: Put a pror on Σ (ν 0 -m-)σ ~ IW(ν 0, (ν 0 -m-)σ 0 ) Step b: Put a pror on Σ Σ ~ N( 0, Σ / κ 0 ) Why do we use ths form of pror? Actually, we don t have to But t s computatonally and algebracally convenent t s a conjugate pror. Copyrght 00, 004, Andrew W. Moore Mamum Lelhood: Slde 45 Beng Bayesan: MAP estmates for Gaussans Suppose you have,, ~(..d) N(,Σ) MAP: Whch (,Σ) mamzes p(,σ,, )? Step : Pror: (ν 0 -m-) Σ ~ IW(ν 0, (ν 0 -m-) Σ 0 ), Σ ~ N( 0, Σ / κ 0 ) Step : κ 0 κ ( ν + m ) Σ ( ν 0 + m ) Σ0 + Copyrght 00, 004, Andrew W. Moore Mamum Lelhood: Slde ν ν 0 + κ κ 0 + T 0 ( )( ) + / κ 0 Step 3: Posteror: (ν +m-)σ ~ IW(ν, (ν +m-) Σ ), Σ ~ N(, Σ / κ ) esult: map, E[Σ,, ] Σ ( )( ) 0 + / T 3

24 Beng Bayesan: Loo MAP carefully estmates what these for formulae Gaussans are Suppose you have,, ~(..d) N(,Σ) MAP: Whch (,Σ) statstcs mamzes of the p(,σ data.,, )? Step : Pror: (ν 0 -m-) Σ ~ The IW(ν margnal 0, (ν 0 -m-) dstrbuton Σ 0 ), Σ on ~ N( s a 0, student-t Σ / κ 0 ) One pont of vew: t s pretty academc f > 30 Step : κ κ + ( ν + m ) Σ ( ν 0 + m ) Σ0 + dong. It s all very sensble. Conjugate prors mean pror form and posteror form are same and characterzed by suffcent Copyrght 00, 004, Andrew W. Moore Mamum Lelhood: Slde 47 0 ν ν 0 + κ κ 0 + T 0 ( )( ) + / κ 0 Step 3: Posteror: (ν +m-)σ ~ IW(ν, (ν +m-) Σ ), Σ ~ N(, Σ / κ ) esult: map, E[Σ,, ] Σ ( )( ) 0 + / T Where we re at Categorcal nputs only eal-valued nputs only Med eal / Cat oay Inputs Classfer Predct category Jont BC Naïve BC Dec Tree Inputs Densty Estmator Probablty Jont DE Naïve DE Gauss DE Inputs egressor Predct real no. Copyrght 00, 004, Andrew W. Moore Mamum Lelhood: Slde 48 4

25 What you should now The ecpe for MLE What do we sometmes prefer MLE to MAP? Understand MLE estmaton of Gaussan parameters Understand based estmator versus unbased estmator Apprecate the outlne behnd Bayesan estmaton of Gaussan parameters Copyrght 00, 004, Andrew W. Moore Mamum Lelhood: Slde 49 Useful eercse We d already done some MLE n ths class wthout even tellng you! Suppose categorcal arty-n nputs,, ~(..d.) from a multnomal M(p, p, p n ) where P( j p)p j What s the MLE p(p, p, p n )? Copyrght 00, 004, Andrew W. Moore Mamum Lelhood: Slde 50 5

Instance-Based Learning (a.k.a. memory-based learning) Part I: Nearest Neighbor Classification

Instance-Based Learning (a.k.a. memory-based learning) Part I: Nearest Neighbor Classification Instance-Based earnng (a.k.a. memory-based learnng) Part I: Nearest Neghbor Classfcaton Note to other teachers and users of these sldes. Andrew would be delghted f you found ths source materal useful n

More information

Clustering with Gaussian Mixtures

Clustering with Gaussian Mixtures Note to other teachers and users of these sldes. Andrew would be delghted f you found ths source materal useful n gvng your own lectures. Feel free to use these sldes verbatm, or to modfy them to ft your

More information

The conjugate prior to a Bernoulli is. A) Bernoulli B) Gaussian C) Beta D) none of the above

The conjugate prior to a Bernoulli is. A) Bernoulli B) Gaussian C) Beta D) none of the above The conjugate pror to a Bernoull s A) Bernoull B) Gaussan C) Beta D) none of the above The conjugate pror to a Gaussan s A) Bernoull B) Gaussan C) Beta D) none of the above MAP estmates A) argmax θ p(θ

More information

Clustering with Gaussian Mixtures

Clustering with Gaussian Mixtures Note to other teachers and users of these sldes. Andrew would be delghted f you found ths source materal useful n gvng your own lectures. Feel free to use these sldes verbatm, or to modfy them to ft your

More information

ENG 8801/ Special Topics in Computer Engineering: Pattern Recognition. Memorial University of Newfoundland Pattern Recognition

ENG 8801/ Special Topics in Computer Engineering: Pattern Recognition. Memorial University of Newfoundland Pattern Recognition EG 880/988 - Specal opcs n Computer Engneerng: Pattern Recognton Memoral Unversty of ewfoundland Pattern Recognton Lecture 7 May 3, 006 http://wwwengrmunca/~charlesr Offce Hours: uesdays hursdays 8:30-9:30

More information

CS 2750 Machine Learning. Lecture 5. Density estimation. CS 2750 Machine Learning. Announcements

CS 2750 Machine Learning. Lecture 5. Density estimation. CS 2750 Machine Learning. Announcements CS 750 Machne Learnng Lecture 5 Densty estmaton Mlos Hauskrecht mlos@cs.ptt.edu 539 Sennott Square CS 750 Machne Learnng Announcements Homework Due on Wednesday before the class Reports: hand n before

More information

Gaussian Mixture Models

Gaussian Mixture Models Lab Gaussan Mxture Models Lab Objectve: Understand the formulaton of Gaussan Mxture Models (GMMs) and how to estmate GMM parameters. You ve already seen GMMs as the observaton dstrbuton n certan contnuous

More information

Expectation Maximization Mixture Models HMMs

Expectation Maximization Mixture Models HMMs -755 Machne Learnng for Sgnal Processng Mture Models HMMs Class 9. 2 Sep 200 Learnng Dstrbutons for Data Problem: Gven a collecton of eamples from some data, estmate ts dstrbuton Basc deas of Mamum Lelhood

More information

MLE and Bayesian Estimation. Jie Tang Department of Computer Science & Technology Tsinghua University 2012

MLE and Bayesian Estimation. Jie Tang Department of Computer Science & Technology Tsinghua University 2012 MLE and Bayesan Estmaton Je Tang Department of Computer Scence & Technology Tsnghua Unversty 01 1 Lnear Regresson? As the frst step, we need to decde how we re gong to represent the functon f. One example:

More information

Using T.O.M to Estimate Parameter of distributions that have not Single Exponential Family

Using T.O.M to Estimate Parameter of distributions that have not Single Exponential Family IOSR Journal of Mathematcs IOSR-JM) ISSN: 2278-5728. Volume 3, Issue 3 Sep-Oct. 202), PP 44-48 www.osrjournals.org Usng T.O.M to Estmate Parameter of dstrbutons that have not Sngle Exponental Famly Jubran

More information

Statistical analysis using matlab. HY 439 Presented by: George Fortetsanakis

Statistical analysis using matlab. HY 439 Presented by: George Fortetsanakis Statstcal analyss usng matlab HY 439 Presented by: George Fortetsanaks Roadmap Probablty dstrbutons Statstcal estmaton Fttng data to probablty dstrbutons Contnuous dstrbutons Contnuous random varable X

More information

xp(x µ) = 0 p(x = 0 µ) + 1 p(x = 1 µ) = µ

xp(x µ) = 0 p(x = 0 µ) + 1 p(x = 1 µ) = µ CSE 455/555 Sprng 2013 Homework 7: Parametrc Technques Jason J. Corso Computer Scence and Engneerng SUY at Buffalo jcorso@buffalo.edu Solutons by Yngbo Zhou Ths assgnment does not need to be submtted and

More information

However, since P is a symmetric idempotent matrix, of P are either 0 or 1 [Eigen-values

However, since P is a symmetric idempotent matrix, of P are either 0 or 1 [Eigen-values Fall 007 Soluton to Mdterm Examnaton STAT 7 Dr. Goel. [0 ponts] For the general lnear model = X + ε, wth uncorrelated errors havng mean zero and varance σ, suppose that the desgn matrx X s not necessarly

More information

Generative classification models

Generative classification models CS 675 Intro to Machne Learnng Lecture Generatve classfcaton models Mlos Hauskrecht mlos@cs.ptt.edu 539 Sennott Square Data: D { d, d,.., dn} d, Classfcaton represents a dscrete class value Goal: learn

More information

MIMA Group. Chapter 2 Bayesian Decision Theory. School of Computer Science and Technology, Shandong University. Xin-Shun SDU

MIMA Group. Chapter 2 Bayesian Decision Theory. School of Computer Science and Technology, Shandong University. Xin-Shun SDU Group M D L M Chapter Bayesan Decson heory Xn-Shun Xu @ SDU School of Computer Scence and echnology, Shandong Unversty Bayesan Decson heory Bayesan decson theory s a statstcal approach to data mnng/pattern

More information

1 GSW Iterative Techniques for y = Ax

1 GSW Iterative Techniques for y = Ax 1 for y = A I m gong to cheat here. here are a lot of teratve technques that can be used to solve the general case of a set of smultaneous equatons (wrtten n the matr form as y = A), but ths chapter sn

More information

10-701/ Machine Learning, Fall 2005 Homework 3

10-701/ Machine Learning, Fall 2005 Homework 3 10-701/15-781 Machne Learnng, Fall 2005 Homework 3 Out: 10/20/05 Due: begnnng of the class 11/01/05 Instructons Contact questons-10701@autonlaborg for queston Problem 1 Regresson and Cross-valdaton [40

More information

Stat 543 Exam 2 Spring 2016

Stat 543 Exam 2 Spring 2016 Stat 543 Exam 2 Sprng 206 I have nether gven nor receved unauthorzed assstance on ths exam. Name Sgned Date Name Prnted Ths Exam conssts of questons. Do at least 0 of the parts of the man exam. I wll score

More information

Hidden Markov Models

Hidden Markov Models Note to other teachers and users of these sldes. Andrew would be delghted f you found ths source materal useful n gvng your own lectures. Feel free to use these sldes verbatm, or to modfy them to ft your

More information

Stat 543 Exam 2 Spring 2016

Stat 543 Exam 2 Spring 2016 Stat 543 Exam 2 Sprng 2016 I have nether gven nor receved unauthorzed assstance on ths exam. Name Sgned Date Name Prnted Ths Exam conssts of 11 questons. Do at least 10 of the 11 parts of the man exam.

More information

Classification learning II

Classification learning II Lecture 8 Classfcaton learnng II Mlos Hauskrecht mlos@cs.ptt.edu 539 Sennott Square Logstc regresson model Defnes a lnear decson boundar Dscrmnant functons: g g g g here g z / e z f, g g - s a logstc functon

More information

Econ107 Applied Econometrics Topic 3: Classical Model (Studenmund, Chapter 4)

Econ107 Applied Econometrics Topic 3: Classical Model (Studenmund, Chapter 4) I. Classcal Assumptons Econ7 Appled Econometrcs Topc 3: Classcal Model (Studenmund, Chapter 4) We have defned OLS and studed some algebrac propertes of OLS. In ths topc we wll study statstcal propertes

More information

MACHINE APPLIED MACHINE LEARNING LEARNING. Gaussian Mixture Regression

MACHINE APPLIED MACHINE LEARNING LEARNING. Gaussian Mixture Regression 11 MACHINE APPLIED MACHINE LEARNING LEARNING MACHINE LEARNING Gaussan Mture Regresson 22 MACHINE APPLIED MACHINE LEARNING LEARNING Bref summary of last week s lecture 33 MACHINE APPLIED MACHINE LEARNING

More information

CIS526: Machine Learning Lecture 3 (Sept 16, 2003) Linear Regression. Preparation help: Xiaoying Huang. x 1 θ 1 output... θ M x M

CIS526: Machine Learning Lecture 3 (Sept 16, 2003) Linear Regression. Preparation help: Xiaoying Huang. x 1 θ 1 output... θ M x M CIS56: achne Learnng Lecture 3 (Sept 6, 003) Preparaton help: Xaoyng Huang Lnear Regresson Lnear regresson can be represented by a functonal form: f(; θ) = θ 0 0 +θ + + θ = θ = 0 ote: 0 s a dummy attrbute

More information

First Year Examination Department of Statistics, University of Florida

First Year Examination Department of Statistics, University of Florida Frst Year Examnaton Department of Statstcs, Unversty of Florda May 7, 010, 8:00 am - 1:00 noon Instructons: 1. You have four hours to answer questons n ths examnaton.. You must show your work to receve

More information

The Gaussian classifier. Nuno Vasconcelos ECE Department, UCSD

The Gaussian classifier. Nuno Vasconcelos ECE Department, UCSD he Gaussan classfer Nuno Vasconcelos ECE Department, UCSD Bayesan decson theory recall that we have state of the world X observatons g decson functon L[g,y] loss of predctng y wth g Bayes decson rule s

More information

1. Inference on Regression Parameters a. Finding Mean, s.d and covariance amongst estimates. 2. Confidence Intervals and Working Hotelling Bands

1. Inference on Regression Parameters a. Finding Mean, s.d and covariance amongst estimates. 2. Confidence Intervals and Working Hotelling Bands Content. Inference on Regresson Parameters a. Fndng Mean, s.d and covarance amongst estmates.. Confdence Intervals and Workng Hotellng Bands 3. Cochran s Theorem 4. General Lnear Testng 5. Measures of

More information

Bayesian Learning. Smart Home Health Analytics Spring Nirmalya Roy Department of Information Systems University of Maryland Baltimore County

Bayesian Learning. Smart Home Health Analytics Spring Nirmalya Roy Department of Information Systems University of Maryland Baltimore County Smart Home Health Analytcs Sprng 2018 Bayesan Learnng Nrmalya Roy Department of Informaton Systems Unversty of Maryland Baltmore ounty www.umbc.edu Bayesan Learnng ombnes pror knowledge wth evdence to

More information

Department of Computer Science Artificial Intelligence Research Laboratory. Iowa State University MACHINE LEARNING

Department of Computer Science Artificial Intelligence Research Laboratory. Iowa State University MACHINE LEARNING MACHINE LEANING Vasant Honavar Bonformatcs and Computatonal Bology rogram Center for Computatonal Intellgence, Learnng, & Dscovery Iowa State Unversty honavar@cs.astate.edu www.cs.astate.edu/~honavar/

More information

Single- Parameter Linear Regression

Single- Parameter Linear Regression Predctng Real-valued outputs an ntroducton to Regresson Note to other teachers and users of these sldes. Andrew would be delghted f ou found ths source materal useful n gvng our own lectures. Feel free

More information

Computation of Higher Order Moments from Two Multinomial Overdispersion Likelihood Models

Computation of Higher Order Moments from Two Multinomial Overdispersion Likelihood Models Computaton of Hgher Order Moments from Two Multnomal Overdsperson Lkelhood Models BY J. T. NEWCOMER, N. K. NEERCHAL Department of Mathematcs and Statstcs, Unversty of Maryland, Baltmore County, Baltmore,

More information

CS 3710: Visual Recognition Classification and Detection. Adriana Kovashka Department of Computer Science January 13, 2015

CS 3710: Visual Recognition Classification and Detection. Adriana Kovashka Department of Computer Science January 13, 2015 CS 3710: Vsual Recognton Classfcaton and Detecton Adrana Kovashka Department of Computer Scence January 13, 2015 Plan for Today Vsual recognton bascs part 2: Classfcaton and detecton Adrana s research

More information

Classification as a Regression Problem

Classification as a Regression Problem Target varable y C C, C,, ; Classfcaton as a Regresson Problem { }, 3 L C K To treat classfcaton as a regresson problem we should transform the target y nto numercal values; The choce of numercal class

More information

AS-Level Maths: Statistics 1 for Edexcel

AS-Level Maths: Statistics 1 for Edexcel 1 of 6 AS-Level Maths: Statstcs 1 for Edecel S1. Calculatng means and standard devatons Ths con ndcates the slde contans actvtes created n Flash. These actvtes are not edtable. For more detaled nstructons,

More information

Fall 2012 Analysis of Experimental Measurements B. Eisenstein/rev. S. Errede

Fall 2012 Analysis of Experimental Measurements B. Eisenstein/rev. S. Errede Fall 0 Analyss of Expermental easurements B. Esensten/rev. S. Errede We now reformulate the lnear Least Squares ethod n more general terms, sutable for (eventually extendng to the non-lnear case, and also

More information

Support Vector Machines

Support Vector Machines CS 2750: Machne Learnng Support Vector Machnes Prof. Adrana Kovashka Unversty of Pttsburgh February 17, 2016 Announcement Homework 2 deadlne s now 2/29 We ll have covered everythng you need today or at

More information

Composite Hypotheses testing

Composite Hypotheses testing Composte ypotheses testng In many hypothess testng problems there are many possble dstrbutons that can occur under each of the hypotheses. The output of the source s a set of parameters (ponts n a parameter

More information

2E Pattern Recognition Solutions to Introduction to Pattern Recognition, Chapter 2: Bayesian pattern classification

2E Pattern Recognition Solutions to Introduction to Pattern Recognition, Chapter 2: Bayesian pattern classification E395 - Pattern Recognton Solutons to Introducton to Pattern Recognton, Chapter : Bayesan pattern classfcaton Preface Ths document s a soluton manual for selected exercses from Introducton to Pattern Recognton

More information

Maximum Likelihood Estimation

Maximum Likelihood Estimation Maxmum Lkelhood Estmaton INFO-2301: Quanttatve Reasonng 2 Mchael Paul and Jordan Boyd-Graber MARCH 7, 2017 INFO-2301: Quanttatve Reasonng 2 Paul and Boyd-Graber Maxmum Lkelhood Estmaton 1 of 9 Why MLE?

More information

Linear Feature Engineering 11

Linear Feature Engineering 11 Lnear Feature Engneerng 11 2 Least-Squares 2.1 Smple least-squares Consder the followng dataset. We have a bunch of nputs x and correspondng outputs y. The partcular values n ths dataset are x y 0.23 0.19

More information

4 Analysis of Variance (ANOVA) 5 ANOVA. 5.1 Introduction. 5.2 Fixed Effects ANOVA

4 Analysis of Variance (ANOVA) 5 ANOVA. 5.1 Introduction. 5.2 Fixed Effects ANOVA 4 Analyss of Varance (ANOVA) 5 ANOVA 51 Introducton ANOVA ANOVA s a way to estmate and test the means of multple populatons We wll start wth one-way ANOVA If the populatons ncluded n the study are selected

More information

Chapter 5. Solution of System of Linear Equations. Module No. 6. Solution of Inconsistent and Ill Conditioned Systems

Chapter 5. Solution of System of Linear Equations. Module No. 6. Solution of Inconsistent and Ill Conditioned Systems Numercal Analyss by Dr. Anta Pal Assstant Professor Department of Mathematcs Natonal Insttute of Technology Durgapur Durgapur-713209 emal: anta.bue@gmal.com 1 . Chapter 5 Soluton of System of Lnear Equatons

More information

Homework Assignment 3 Due in class, Thursday October 15

Homework Assignment 3 Due in class, Thursday October 15 Homework Assgnment 3 Due n class, Thursday October 15 SDS 383C Statstcal Modelng I 1 Rdge regresson and Lasso 1. Get the Prostrate cancer data from http://statweb.stanford.edu/~tbs/elemstatlearn/ datasets/prostate.data.

More information

Parametric fractional imputation for missing data analysis. Jae Kwang Kim Survey Working Group Seminar March 29, 2010

Parametric fractional imputation for missing data analysis. Jae Kwang Kim Survey Working Group Seminar March 29, 2010 Parametrc fractonal mputaton for mssng data analyss Jae Kwang Km Survey Workng Group Semnar March 29, 2010 1 Outlne Introducton Proposed method Fractonal mputaton Approxmaton Varance estmaton Multple mputaton

More information

Which Separator? Spring 1

Which Separator? Spring 1 Whch Separator? 6.034 - Sprng 1 Whch Separator? Mamze the margn to closest ponts 6.034 - Sprng Whch Separator? Mamze the margn to closest ponts 6.034 - Sprng 3 Margn of a pont " # y (w $ + b) proportonal

More information

Why Bayesian? 3. Bayes and Normal Models. State of nature: class. Decision rule. Rev. Thomas Bayes ( ) Bayes Theorem (yes, the famous one)

Why Bayesian? 3. Bayes and Normal Models. State of nature: class. Decision rule. Rev. Thomas Bayes ( ) Bayes Theorem (yes, the famous one) Why Bayesan? 3. Bayes and Normal Models Alex M. Martnez alex@ece.osu.edu Handouts Handoutsfor forece ECE874 874Sp Sp007 If all our research (n PR was to dsappear and you could only save one theory, whch

More information

3.1 Expectation of Functions of Several Random Variables. )' be a k-dimensional discrete or continuous random vector, with joint PMF p (, E X E X1 E X

3.1 Expectation of Functions of Several Random Variables. )' be a k-dimensional discrete or continuous random vector, with joint PMF p (, E X E X1 E X Statstcs 1: Probablty Theory II 37 3 EPECTATION OF SEVERAL RANDOM VARIABLES As n Probablty Theory I, the nterest n most stuatons les not on the actual dstrbuton of a random vector, but rather on a number

More information

Conjugacy and the Exponential Family

Conjugacy and the Exponential Family CS281B/Stat241B: Advanced Topcs n Learnng & Decson Makng Conjugacy and the Exponental Famly Lecturer: Mchael I. Jordan Scrbes: Bran Mlch 1 Conjugacy In the prevous lecture, we saw conjugate prors for the

More information

Differentiating Gaussian Processes

Differentiating Gaussian Processes Dfferentatng Gaussan Processes Andrew McHutchon Aprl 17, 013 1 Frst Order Dervatve of the Posteror Mean The posteror mean of a GP s gven by, f = x, X KX, X 1 y x, X α 1 Only the x, X term depends on the

More information

Logistic Regression. CAP 5610: Machine Learning Instructor: Guo-Jun QI

Logistic Regression. CAP 5610: Machine Learning Instructor: Guo-Jun QI Logstc Regresson CAP 561: achne Learnng Instructor: Guo-Jun QI Bayes Classfer: A Generatve model odel the posteror dstrbuton P(Y X) Estmate class-condtonal dstrbuton P(X Y) for each Y Estmate pror dstrbuton

More information

Logistic Regression Maximum Likelihood Estimation

Logistic Regression Maximum Likelihood Estimation Harvard-MIT Dvson of Health Scences and Technology HST.951J: Medcal Decson Support, Fall 2005 Instructors: Professor Lucla Ohno-Machado and Professor Staal Vnterbo 6.873/HST.951 Medcal Decson Support Fall

More information

Department of Statistics University of Toronto STA305H1S / 1004 HS Design and Analysis of Experiments Term Test - Winter Solution

Department of Statistics University of Toronto STA305H1S / 1004 HS Design and Analysis of Experiments Term Test - Winter Solution Department of Statstcs Unversty of Toronto STA35HS / HS Desgn and Analyss of Experments Term Test - Wnter - Soluton February, Last Name: Frst Name: Student Number: Instructons: Tme: hours. Ads: a non-programmable

More information

Primer on High-Order Moment Estimators

Primer on High-Order Moment Estimators Prmer on Hgh-Order Moment Estmators Ton M. Whted July 2007 The Errors-n-Varables Model We wll start wth the classcal EIV for one msmeasured regressor. The general case s n Erckson and Whted Econometrc

More information

Chapter 14 Simple Linear Regression

Chapter 14 Simple Linear Regression Chapter 4 Smple Lnear Regresson Chapter 4 - Smple Lnear Regresson Manageral decsons often are based on the relatonshp between two or more varables. Regresson analss can be used to develop an equaton showng

More information

Issues To Consider when Estimating Health Care Costs with Generalized Linear Models (GLMs): To Gamma/Log Or Not To Gamma/Log? That Is The New Question

Issues To Consider when Estimating Health Care Costs with Generalized Linear Models (GLMs): To Gamma/Log Or Not To Gamma/Log? That Is The New Question Issues To Consder when Estmatng Health Care Costs wth Generalzed Lnear Models (GLMs): To Gamma/Log Or Not To Gamma/Log? That Is The New Queston ISPOR 20th Annual Internatonal Meetng May 19, 2015 Jalpa

More information

Image classification. Given the bag-of-features representations of images from different classes, how do we learn a model for distinguishing i them?

Image classification. Given the bag-of-features representations of images from different classes, how do we learn a model for distinguishing i them? Image classfcaton Gven te bag-of-features representatons of mages from dfferent classes ow do we learn a model for dstngusng tem? Classfers Learn a decson rule assgnng bag-offeatures representatons of

More information

Comparison of Regression Lines

Comparison of Regression Lines STATGRAPHICS Rev. 9/13/2013 Comparson of Regresson Lnes Summary... 1 Data Input... 3 Analyss Summary... 4 Plot of Ftted Model... 6 Condtonal Sums of Squares... 6 Analyss Optons... 7 Forecasts... 8 Confdence

More information

Learning from Data 1 Naive Bayes

Learning from Data 1 Naive Bayes Learnng from Data 1 Nave Bayes Davd Barber dbarber@anc.ed.ac.uk course page : http://anc.ed.ac.uk/ dbarber/lfd1/lfd1.html c Davd Barber 2001, 2002 1 Learnng from Data 1 : c Davd Barber 2001,2002 2 1 Why

More information

Properties of Least Squares

Properties of Least Squares Week 3 3.1 Smple Lnear Regresson Model 3. Propertes of Least Squares Estmators Y Y β 1 + β X + u weekly famly expendtures X weekly famly ncome For a gven level of x, the expected level of food expendtures

More information

UNIVERSITY OF TORONTO Faculty of Arts and Science. December 2005 Examinations STA437H1F/STA1005HF. Duration - 3 hours

UNIVERSITY OF TORONTO Faculty of Arts and Science. December 2005 Examinations STA437H1F/STA1005HF. Duration - 3 hours UNIVERSITY OF TORONTO Faculty of Arts and Scence December 005 Examnatons STA47HF/STA005HF Duraton - hours AIDS ALLOWED: (to be suppled by the student) Non-programmable calculator One handwrtten 8.5'' x

More information

Mixture o f of Gaussian Gaussian clustering Nov

Mixture o f of Gaussian Gaussian clustering Nov Mture of Gaussan clusterng Nov 11 2009 Soft vs hard lusterng Kmeans performs Hard clusterng: Data pont s determnstcally assgned to one and only one cluster But n realty clusters may overlap Soft-clusterng:

More information

β0 + β1xi and want to estimate the unknown

β0 + β1xi and want to estimate the unknown SLR Models Estmaton Those OLS Estmates Estmators (e ante) v. estmates (e post) The Smple Lnear Regresson (SLR) Condtons -4 An Asde: The Populaton Regresson Functon B and B are Lnear Estmators (condtonal

More information

Statistical Foundations of Pattern Recognition

Statistical Foundations of Pattern Recognition Statstcal Foundatons of Pattern Recognton Learnng Objectves Bayes Theorem Decson-mang Confdence factors Dscrmnants The connecton to neural nets Statstcal Foundatons of Pattern Recognton NDE measurement

More information

Department of Quantitative Methods & Information Systems. Time Series and Their Components QMIS 320. Chapter 6

Department of Quantitative Methods & Information Systems. Time Series and Their Components QMIS 320. Chapter 6 Department of Quanttatve Methods & Informaton Systems Tme Seres and Ther Components QMIS 30 Chapter 6 Fall 00 Dr. Mohammad Zanal These sldes were modfed from ther orgnal source for educatonal purpose only.

More information

Outline. Multivariate Parametric Methods. Multivariate Data. Basic Multivariate Statistics. Steven J Zeil

Outline. Multivariate Parametric Methods. Multivariate Data. Basic Multivariate Statistics. Steven J Zeil Outlne Multvarate Parametrc Methods Steven J Zel Old Domnon Unv. Fall 2010 1 Multvarate Data 2 Multvarate ormal Dstrbuton 3 Multvarate Classfcaton Dscrmnants Tunng Complexty Dscrete Features 4 Multvarate

More information

LINEAR REGRESSION ANALYSIS. MODULE IX Lecture Multicollinearity

LINEAR REGRESSION ANALYSIS. MODULE IX Lecture Multicollinearity LINEAR REGRESSION ANALYSIS MODULE IX Lecture - 30 Multcollnearty Dr. Shalabh Department of Mathematcs and Statstcs Indan Insttute of Technology Kanpur 2 Remedes for multcollnearty Varous technques have

More information

Lecture 6: Introduction to Linear Regression

Lecture 6: Introduction to Linear Regression Lecture 6: Introducton to Lnear Regresson An Manchakul amancha@jhsph.edu 24 Aprl 27 Lnear regresson: man dea Lnear regresson can be used to study an outcome as a lnear functon of a predctor Example: 6

More information

MATH 829: Introduction to Data Mining and Analysis The EM algorithm (part 2)

MATH 829: Introduction to Data Mining and Analysis The EM algorithm (part 2) 1/16 MATH 829: Introducton to Data Mnng and Analyss The EM algorthm (part 2) Domnque Gullot Departments of Mathematcal Scences Unversty of Delaware Aprl 20, 2016 Recall 2/16 We are gven ndependent observatons

More information

Randomness and Computation

Randomness and Computation Randomness and Computaton or, Randomzed Algorthms Mary Cryan School of Informatcs Unversty of Ednburgh RC 208/9) Lecture 0 slde Balls n Bns m balls, n bns, and balls thrown unformly at random nto bns usually

More information

Lecture 12: Classification

Lecture 12: Classification Lecture : Classfcaton g Dscrmnant functons g The optmal Bayes classfer g Quadratc classfers g Eucldean and Mahalanobs metrcs g K Nearest Neghbor Classfers Intellgent Sensor Systems Rcardo Guterrez-Osuna

More information

Chapter 2 - The Simple Linear Regression Model S =0. e i is a random error. S β2 β. This is a minimization problem. Solution is a calculus exercise.

Chapter 2 - The Simple Linear Regression Model S =0. e i is a random error. S β2 β. This is a minimization problem. Solution is a calculus exercise. Chapter - The Smple Lnear Regresson Model The lnear regresson equaton s: where y + = β + β e for =,..., y and are observable varables e s a random error How can an estmaton rule be constructed for the

More information

e i is a random error

e i is a random error Chapter - The Smple Lnear Regresson Model The lnear regresson equaton s: where + β + β e for,..., and are observable varables e s a random error How can an estmaton rule be constructed for the unknown

More information

Basically, if you have a dummy dependent variable you will be estimating a probability.

Basically, if you have a dummy dependent variable you will be estimating a probability. ECON 497: Lecture Notes 13 Page 1 of 1 Metropoltan State Unversty ECON 497: Research and Forecastng Lecture Notes 13 Dummy Dependent Varable Technques Studenmund Chapter 13 Bascally, f you have a dummy

More information

An Experiment/Some Intuition (Fall 2006): Lecture 18 The EM Algorithm heads coin 1 tails coin 2 Overview Maximum Likelihood Estimation

An Experiment/Some Intuition (Fall 2006): Lecture 18 The EM Algorithm heads coin 1 tails coin 2 Overview Maximum Likelihood Estimation An Experment/Some Intuton I have three cons n my pocket, 6.864 (Fall 2006): Lecture 18 The EM Algorthm Con 0 has probablty λ of heads; Con 1 has probablty p 1 of heads; Con 2 has probablty p 2 of heads

More information

Linear regression. Regression Models. Chapter 11 Student Lecture Notes Regression Analysis is the

Linear regression. Regression Models. Chapter 11 Student Lecture Notes Regression Analysis is the Chapter 11 Student Lecture Notes 11-1 Lnear regresson Wenl lu Dept. Health statstcs School of publc health Tanjn medcal unversty 1 Regresson Models 1. Answer What Is the Relatonshp Between the Varables?.

More information

j) = 1 (note sigma notation) ii. Continuous random variable (e.g. Normal distribution) 1. density function: f ( x) 0 and f ( x) dx = 1

j) = 1 (note sigma notation) ii. Continuous random variable (e.g. Normal distribution) 1. density function: f ( x) 0 and f ( x) dx = 1 Random varables Measure of central tendences and varablty (means and varances) Jont densty functons and ndependence Measures of assocaton (covarance and correlaton) Interestng result Condtonal dstrbutons

More information

+, where 0 x N - n. k k

+, where 0 x N - n. k k CO 745, Mdterm Len Cabrera. A multle choce eam has questons, each of whch has ossble answers. A student nows the correct answer to n of these questons. For the remanng - n questons, he checs the answers

More information

Predictive Analytics : QM901.1x Prof U Dinesh Kumar, IIMB. All Rights Reserved, Indian Institute of Management Bangalore

Predictive Analytics : QM901.1x Prof U Dinesh Kumar, IIMB. All Rights Reserved, Indian Institute of Management Bangalore Sesson Outlne Introducton to classfcaton problems and dscrete choce models. Introducton to Logstcs Regresson. Logstc functon and Logt functon. Maxmum Lkelhood Estmator (MLE) for estmaton of LR parameters.

More information

CHAPTER 3: BAYESIAN DECISION THEORY

CHAPTER 3: BAYESIAN DECISION THEORY HATER 3: BAYESIAN DEISION THEORY Decson mang under uncertanty 3 Data comes from a process that s completely not nown The lac of nowledge can be compensated by modelng t as a random process May be the underlyng

More information

Integrals and Invariants of Euler-Lagrange Equations

Integrals and Invariants of Euler-Lagrange Equations Lecture 16 Integrals and Invarants of Euler-Lagrange Equatons ME 256 at the Indan Insttute of Scence, Bengaluru Varatonal Methods and Structural Optmzaton G. K. Ananthasuresh Professor, Mechancal Engneerng,

More information

Probability Density Function Estimation by different Methods

Probability Density Function Estimation by different Methods EEE 739Q SPRIG 00 COURSE ASSIGMET REPORT Probablty Densty Functon Estmaton by dfferent Methods Vas Chandraant Rayar Abstract The am of the assgnment was to estmate the probablty densty functon (PDF of

More information

SDMML HT MSc Problem Sheet 4

SDMML HT MSc Problem Sheet 4 SDMML HT 06 - MSc Problem Sheet 4. The recever operatng characterstc ROC curve plots the senstvty aganst the specfcty of a bnary classfer as the threshold for dscrmnaton s vared. Let the data space be

More information

Logistic Classifier CISC 5800 Professor Daniel Leeds

Logistic Classifier CISC 5800 Professor Daniel Leeds lon 9/7/8 Logstc Classfer CISC 58 Professor Danel Leeds Classfcaton strategy: generatve vs. dscrmnatve Generatve, e.g., Bayes/Naïve Bayes: 5 5 Identfy probablty dstrbuton for each class Determne class

More information

Lecture 9: Linear regression: centering, hypothesis testing, multiple covariates, and confounding

Lecture 9: Linear regression: centering, hypothesis testing, multiple covariates, and confounding Recall: man dea of lnear regresson Lecture 9: Lnear regresson: centerng, hypothess testng, multple covarates, and confoundng Sandy Eckel seckel@jhsph.edu 6 May 8 Lnear regresson can be used to study an

More information

Lecture 9: Linear regression: centering, hypothesis testing, multiple covariates, and confounding

Lecture 9: Linear regression: centering, hypothesis testing, multiple covariates, and confounding Lecture 9: Lnear regresson: centerng, hypothess testng, multple covarates, and confoundng Sandy Eckel seckel@jhsph.edu 6 May 008 Recall: man dea of lnear regresson Lnear regresson can be used to study

More information

Bayesian predictive Configural Frequency Analysis

Bayesian predictive Configural Frequency Analysis Psychologcal Test and Assessment Modelng, Volume 54, 2012 (3), 285-292 Bayesan predctve Confgural Frequency Analyss Eduardo Gutérrez-Peña 1 Abstract Confgural Frequency Analyss s a method for cell-wse

More information

Machine learning: Density estimation

Machine learning: Density estimation CS 70 Foundatons of AI Lecture 3 Machne learnng: ensty estmaton Mlos Hauskrecht mlos@cs.ptt.edu 539 Sennott Square ata: ensty estmaton {.. n} x a vector of attrbute values Objectve: estmate the model of

More information

Statistics Chapter 4

Statistics Chapter 4 Statstcs Chapter 4 "There are three knds of les: les, damned les, and statstcs." Benjamn Dsrael, 1895 (Brtsh statesman) Gaussan Dstrbuton, 4-1 If a measurement s repeated many tmes a statstcal treatment

More information

x i1 =1 for all i (the constant ).

x i1 =1 for all i (the constant ). Chapter 5 The Multple Regresson Model Consder an economc model where the dependent varable s a functon of K explanatory varables. The economc model has the form: y = f ( x,x,..., ) xk Approxmate ths by

More information

Hidden Markov Models & The Multivariate Gaussian (10/26/04)

Hidden Markov Models & The Multivariate Gaussian (10/26/04) CS281A/Stat241A: Statstcal Learnng Theory Hdden Markov Models & The Multvarate Gaussan (10/26/04) Lecturer: Mchael I. Jordan Scrbes: Jonathan W. Hu 1 Hdden Markov Models As a bref revew, hdden Markov models

More information

Grover s Algorithm + Quantum Zeno Effect + Vaidman

Grover s Algorithm + Quantum Zeno Effect + Vaidman Grover s Algorthm + Quantum Zeno Effect + Vadman CS 294-2 Bomb 10/12/04 Fall 2004 Lecture 11 Grover s algorthm Recall that Grover s algorthm for searchng over a space of sze wors as follows: consder the

More information

Stat 642, Lecture notes for 01/27/ d i = 1 t. n i t nj. n j

Stat 642, Lecture notes for 01/27/ d i = 1 t. n i t nj. n j Stat 642, Lecture notes for 01/27/05 18 Rate Standardzaton Contnued: Note that f T n t where T s the cumulatve follow-up tme and n s the number of subjects at rsk at the mdpont or nterval, and d s the

More information

Statistics for Business and Economics

Statistics for Business and Economics Statstcs for Busness and Economcs Chapter 11 Smple Regresson Copyrght 010 Pearson Educaton, Inc. Publshng as Prentce Hall Ch. 11-1 11.1 Overvew of Lnear Models n An equaton can be ft to show the best lnear

More information

Statistics MINITAB - Lab 2

Statistics MINITAB - Lab 2 Statstcs 20080 MINITAB - Lab 2 1. Smple Lnear Regresson In smple lnear regresson we attempt to model a lnear relatonshp between two varables wth a straght lne and make statstcal nferences concernng that

More information

The EM Algorithm (Dempster, Laird, Rubin 1977) The missing data or incomplete data setting: ODL(φ;Y ) = [Y;φ] = [Y X,φ][X φ] = X

The EM Algorithm (Dempster, Laird, Rubin 1977) The missing data or incomplete data setting: ODL(φ;Y ) = [Y;φ] = [Y X,φ][X φ] = X The EM Algorthm (Dempster, Lard, Rubn 1977 The mssng data or ncomplete data settng: An Observed Data Lkelhood (ODL that s a mxture or ntegral of Complete Data Lkelhoods (CDL. (1a ODL(;Y = [Y;] = [Y,][

More information

Pattern Classification

Pattern Classification Pattern Classfcaton All materals n these sldes ere taken from Pattern Classfcaton (nd ed) by R. O. Duda, P. E. Hart and D. G. Stork, John Wley & Sons, 000 th the permsson of the authors and the publsher

More information

Lecture 10 Support Vector Machines II

Lecture 10 Support Vector Machines II Lecture 10 Support Vector Machnes II 22 February 2016 Taylor B. Arnold Yale Statstcs STAT 365/665 1/28 Notes: Problem 3 s posted and due ths upcomng Frday There was an early bug n the fake-test data; fxed

More information

THE ROYAL STATISTICAL SOCIETY 2006 EXAMINATIONS SOLUTIONS HIGHER CERTIFICATE

THE ROYAL STATISTICAL SOCIETY 2006 EXAMINATIONS SOLUTIONS HIGHER CERTIFICATE THE ROYAL STATISTICAL SOCIETY 6 EXAMINATIONS SOLUTIONS HIGHER CERTIFICATE PAPER I STATISTICAL THEORY The Socety provdes these solutons to assst canddates preparng for the eamnatons n future years and for

More information

Limited Dependent Variables

Limited Dependent Variables Lmted Dependent Varables. What f the left-hand sde varable s not a contnuous thng spread from mnus nfnty to plus nfnty? That s, gven a model = f (, β, ε, where a. s bounded below at zero, such as wages

More information

Space of ML Problems. CSE 473: Artificial Intelligence. Parameter Estimation and Bayesian Networks. Learning Topics

Space of ML Problems. CSE 473: Artificial Intelligence. Parameter Estimation and Bayesian Networks. Learning Topics /7/7 CSE 73: Artfcal Intellgence Bayesan - Learnng Deter Fox Sldes adapted from Dan Weld, Jack Breese, Dan Klen, Daphne Koller, Stuart Russell, Andrew Moore & Luke Zettlemoyer What s Beng Learned? Space

More information