Learning with Maximum Likelihood
|
|
- Amber Ethel Gibson
- 6 years ago
- Views:
Transcription
1 Learnng wth Mamum Lelhood Note to other teachers and users of these sldes. Andrew would be delghted f you found ths source materal useful n gvng your own lectures. Feel free to use these sldes verbatm, or to modfy them to ft your own needs. PowerPont orgnals are avalable. If you mae use of a sgnfcant porton of these sldes n your own lecture, please nclude ths message, or the followng ln to the source repostory of Andrew s tutorals: Comments and correctons gratefully receved. Andrew W. Moore Professor School of Computer Scence Carnege Mellon Unversty awm@cs.cmu.edu Copyrght 00, 004, Andrew W. Moore Sep 6th, 00 Mamum Lelhood learnng of Gaussans for Data Mnng Why we should care Learnng Unvarate Gaussans Learnng Multvarate Gaussans What s a based estmator? Bayesan Learnng of Gaussans Copyrght 00, 004, Andrew W. Moore Mamum Lelhood: Slde
2 Why we should care Mamum Lelhood Estmaton s a very very very very fundamental part of data analyss. MLE for Gaussans s tranng wheels for our future technques Learnng Gaussans s more useful than you mght guess Copyrght 00, 004, Andrew W. Moore Mamum Lelhood: Slde 3 Learnng Gaussans from Data Suppose you have,, ~ (..d) N(, ) But you don t now (you do now ) MLE: For whch s,, most lely? MAP: Whch mamzes p(,,, )? Copyrght 00, 004, Andrew W. Moore Mamum Lelhood: Slde 4
3 Learnng Gaussans from Data Suppose you have,, ~(..d) N(, ) But you don t now Sneer (you do now ) MLE: For whch s,, most lely? MAP: Whch mamzes p(,,, )? Copyrght 00, 004, Andrew W. Moore Mamum Lelhood: Slde 5 Learnng Gaussans from Data Suppose you have,, ~(..d) N(, ) But you don t now Sneer (you do now ) MLE: For whch s,, most lely? MAP: Whch mamzes p(,,, )? Despte ths, we ll spend 95% of our tme on MLE. Why? Wat and see Copyrght 00, 004, Andrew W. Moore Mamum Lelhood: Slde 6 3
4 MLE for unvarate Gaussan Suppose you have,, ~(..d) N(, ) But you don t now (you do now ) MLE: For whch s,, most lely? arg ma p(,,..., ) Copyrght 00, 004, Andrew W. Moore Mamum Lelhood: Slde 7 Algebra Euphora arg ma p(,,..., ) (by..d) (monotoncty of log) (plug n formula for Gaussan) (after smplfcaton) Copyrght 00, 004, Andrew W. Moore Mamum Lelhood: Slde 8 4
5 arg ma p( Algebra Euphora arg ma p(,,..., ) arg ma arg ma log p( π, ( arg mn ( ), ) ) ) (by..d) (monotoncty of log) (plug n formula for Gaussan) (after smplfcaton) Copyrght 00, 004, Andrew W. Moore Mamum Lelhood: Slde 9 Intermsson: A General Scalar MLE strategy Tas: Fnd MLE θ assumng nown form for p(data θ,stuff). Wrte LL log P(Data θ,stuff). Wor out / θ usng hgh-school calculus 3. Set / θ0 for a mamum, creatng an equaton n terms of θ 4. Solve t* 5. Chec that you ve found a mamum rather than a mnmum or saddle-pont, and be careful f θ s constraned *Ths s a perfect eample of somethng that wors perfectly n all tetboo eamples and usually nvolves surprsng pan f you need t for somethng new. Copyrght 00, 004, Andrew W. Moore Mamum Lelhood: Slde 0 5
6 The MLE arg ma p(,,..., ) arg mn ( ) s.t. 0 (what?) Copyrght 00, 004, Andrew W. Moore Mamum Lelhood: Slde The MLE arg ma p(,,..., ) arg mn ( ) s.t. 0 Thus Copyrght 00, 004, Andrew W. Moore Mamum Lelhood: Slde ( ) ( ) 6
7 Laws-a-lawdy! The best estmate of the mean of a dstrbuton s the mean of the sample! At frst sght: Ths nd of pedantc, algebra-flled and ultmately unsurprsng fact s eactly the reason people throw down ther Statstcs boo and pc up ther Agent Based Evolutonary Data Mnng Usng The Neuro-Fuzz Transform boo. Copyrght 00, 004, Andrew W. Moore Mamum Lelhood: Slde 3 A General MLE strategy Suppose θ (θ, θ,, θ n ) T s a vector of parameters. Tas: Fnd MLE θ assumng nown form for p(data θ,stuff). Wrte LL log P(Data θ,stuff). Wor out / θ usng hgh-school calculus θ θ θ M θ n Copyrght 00, 004, Andrew W. Moore Mamum Lelhood: Slde 4 7
8 A General MLE strategy Suppose θ (θ, θ,, θ n ) T s a vector of parameters. Tas: Fnd MLE θ assumng nown form for p(data θ,stuff). Wrte LL log P(Data θ,stuff). Wor out / θ usng hgh-school calculus 3. Solve the set of smultaneous equatons θ θ M θ n Copyrght 00, 004, Andrew W. Moore Mamum Lelhood: Slde 5 A General MLE strategy Suppose θ (θ, θ,, θ n ) T s a vector of parameters. Tas: Fnd MLE θ assumng nown form for p(data θ,stuff). Wrte LL log P(Data θ,stuff). Wor out / θ usng hgh-school calculus 3. Solve the set of smultaneous equatons 0 θ 0 θ M 0 θ n 4. Chec that you re at a mamum Copyrght 00, 004, Andrew W. Moore Mamum Lelhood: Slde 6 8
9 A General MLE strategy Suppose θ (θ, θ,, θ n ) T s a vector of parameters. Tas: Fnd MLE θ assumng nown form for p(data θ,stuff). Wrte LL log P(Data θ,stuff). Wor out / θ usng hgh-school calculus 3. Solve the set of smultaneous equatons If you can t solve them, what should you do? 0 θ 0 θ M 0 θ n 4. Chec that you re at a mamum Copyrght 00, 004, Andrew W. Moore Mamum Lelhood: Slde 7 MLE for unvarate Gaussan Suppose you have,, ~(..d) N(, ) But you don t now or MLE: For whch θ (, ) s,, most lely? log p(,,..., ) (log π + log ) ( ) ( ) + ( ) 4 Copyrght 00, 004, Andrew W. Moore Mamum Lelhood: Slde 8 9
10 MLE for unvarate Gaussan Suppose you have,, ~(..d) N(, ) But you don t now or MLE: For whch θ (, ) s,, most lely? log p(,,..., ) (log π + log ) ( ) 0 ( ) 0 + ( ) 4 Copyrght 00, 004, Andrew W. Moore Mamum Lelhood: Slde 9 MLE for unvarate Gaussan Suppose you have,, ~(..d) N(, ) But you don t now or MLE: For whch θ (, ) s,, most lely? log p(,,..., ) (log π + log ) ( ) 0 ( ) 0 + ( ) 4 what? Copyrght 00, 004, Andrew W. Moore Mamum Lelhood: Slde 0 0
11 MLE for unvarate Gaussan Suppose you have,, ~(..d) N(, ) But you don t now or MLE: For whch θ (, ) s,, most lely? ( ) Copyrght 00, 004, Andrew W. Moore Mamum Lelhood: Slde Unbased Estmators An estmator of a parameter s unbased f the epected value of the estmate s the same as the true value of the parameters. If,, ~(..d) N(, ) then E[ ] E s unbased Copyrght 00, 004, Andrew W. Moore Mamum Lelhood: Slde
12 Copyrght 00, 004, Andrew W. Moore Mamum Lelhood: Slde 3 Based Estmators An estmator of a parameter s based f the epected value of the estmate s dfferent from the true value of the parameters. If,, ~(..d) N(, ) then [ ] ) ( j j E E E s based Copyrght 00, 004, Andrew W. Moore Mamum Lelhood: Slde 4 MLE Varance Bas If,, ~(..d) N(, ) then [ ] E E j j Intuton chec: consder the case of Why should our guts epect that would be an underestmate of true? How could you prove that?
13 Unbased estmate of Varance If,, ~(..d) N(, ) then E [ ] E j j So defne unbased [ ] So E unbased Copyrght 00, 004, Andrew W. Moore Mamum Lelhood: Slde 5 Unbased estmate of Varance If,, ~(..d) N(, ) then E [ ] E j j So defne unbased [ ] So E unbased unbased ( ) Copyrght 00, 004, Andrew W. Moore Mamum Lelhood: Slde 6 3
14 Unbasedtude dscusson Whch s best? ( unbased ( ) ) Answer: It depends on the tas And doesn t mae much dfference once --> large Copyrght 00, 004, Andrew W. Moore Mamum Lelhood: Slde 7 Don t get too ected about beng unbased Assume,, ~(..d) N(, ) Suppose we had these estmators for the mean suboptmal + 7 crap Are ether of these unbased? Wll ether of them asymptote to the correct value as gets large? Whch s more useful? Copyrght 00, 004, Andrew W. Moore Mamum Lelhood: Slde 8 4
15 MLE for m-dmensonal Gaussan Suppose you have,, ~(..d) N(,Σ) But you don t now or Σ MLE: For whch θ (,Σ) s,, most lely? Σ ( )( ) T Copyrght 00, 004, Andrew W. Moore Mamum Lelhood: Slde 9 MLE for m-dmensonal Gaussan Suppose you have,, ~(..d) N(,Σ) But you don t now or Σ MLE: For whch θ (,Σ) s,, most lely? Σ ( )( ) T Where m And s value of the th component of (the th attrbute of the th record) And s the th component of Copyrght 00, 004, Andrew W. Moore Mamum Lelhood: Slde 30 5
16 MLE for m-dmensonal Gaussan Suppose you have,, ~(..d) N(,Σ) But you don t now or Σ MLE: For whch θ (,Σ) s,, most lely? Σ ( )( ) j Copyrght 00, 004, Andrew W. Moore Mamum Lelhood: Slde 3 T ( )( j ) j Where m, j m And s value of the th component of (the th attrbute of the th record) And j s the (,j) th component of Σ MLE for m-dmensonal Gaussan Suppose you have,, ~(..d) N(,Σ) But you don t now or Σ MLE: For whch θ (,Σ) s Note, how, Σ most s forced lely? to be Σ ( )( ) Σ unbased Σ Copyrght 00, 004, Andrew W. Moore Mamum Lelhood: Slde 3 T Q: How would you prove ths? A: Just plug through the MLE recpe. symmetrc non-negatve defnte Note the unbased case How many dataponts would you need before the Gaussan has a chance of beng non-degenerate? ( )( ) T 6
17 Confdence ntervals We need to tal We need to dscuss how accurate we epect and Σ to be as a functon of And we need to consder how to estmate these accuraces from data Analytcally * Non-parametrcally (usng randomzaton and bootstrappng) * But we won t. Not yet. *Wll be dscussed n future Andrew lectures just before we need ths technology. Copyrght 00, 004, Andrew W. Moore Mamum Lelhood: Slde 33 Structural error Actually, we need to tal about somethng else too.. What f we do all ths analyss when the true dstrbuton s n fact not Gaussan? How can we tell? * How can we survve? * *Wll be dscussed n future Andrew lectures just before we need ths technology. Copyrght 00, 004, Andrew W. Moore Mamum Lelhood: Slde 34 7
18 Gaussan MLE n acton Usng 39 cars from the MPG UCI dataset suppled by oss Qunlan Copyrght 00, 004, Andrew W. Moore Mamum Lelhood: Slde 35 Data-starved Gaussan MLE Usng three subsets of MPG. Each subset has 6 randomly-chosen cars. Copyrght 00, 004, Andrew W. Moore Mamum Lelhood: Slde 36 8
19 Bvarate MLE n acton Copyrght 00, 004, Andrew W. Moore Mamum Lelhood: Slde 37 Multvarate MLE Covarance matrces are not ectng to loo at Copyrght 00, 004, Andrew W. Moore Mamum Lelhood: Slde 38 9
20 Beng Bayesan: MAP estmates for Gaussans Suppose you have,, ~(..d) N(,Σ) But you don t now or Σ MAP: Whch (,Σ) mamzes p(,σ,, )? Step : Put a pror on (,Σ) Copyrght 00, 004, Andrew W. Moore Mamum Lelhood: Slde 39 Beng Bayesan: MAP estmates for Gaussans Suppose you have,, ~(..d) N(,Σ) But you don t now or Σ MAP: Whch (,Σ) mamzes p(,σ,, )? Step : Put a pror on (,Σ) Step a: Put a pror on Σ (ν 0 -m-) Σ ~ IW(ν 0, (ν 0 -m-) Σ 0 ) Ths thng s called the Inverse-Wshart dstrbuton. A PDF over SPD matrces! Copyrght 00, 004, Andrew W. Moore Mamum Lelhood: Slde 40 0
21 Beng Bayesan: MAP estmates for Gaussans ν Suppose 0 small: I am not sure you have,, ~(..d) N(,Σ) about my guess of Σ 0 Σ But you don t now or Σ 0 : (oughly) my best guess of Σ MAP: ν 0 large: Whch I m pretty (,Σ) sure mamzes p(,σ,, )? about my guess of Σ 0 Ε[Σ ] Σ 0 Step : Put a pror on (,Σ) Step a: Put a pror on Σ (ν 0 -m-) Σ ~ IW(ν 0, (ν 0 -m-) Σ 0 ) Ths thng s called the Inverse-Wshart dstrbuton. A PDF over SPD matrces! Copyrght 00, 004, Andrew W. Moore Mamum Lelhood: Slde 4 Beng Bayesan: MAP estmates for Gaussans Suppose you have,, ~(..d) N(,Σ) But you don t now or Σ MAP: Whch (,Σ) mamzes p(,σ,, )? Step : Put a pror on (,Σ) Step a: Put a pror on Σ (ν 0 -m-)σ ~ IW(ν 0, (ν 0 -m-)σ 0 ) Step b: Put a pror on Σ Σ ~ N( 0, Σ / κ 0 ) Together, Σ and Σ defne a jont dstrbuton on (,Σ) Copyrght 00, 004, Andrew W. Moore Mamum Lelhood: Slde 4
22 Beng Bayesan: MAP estmates for Gaussans Suppose you have,, ~(..d) N(,Σ) But you don t now or Σ κ 0 small: I am not sure MAP: Whch (,Σ) mamzes about p(,σ my guess, of, 0 )? 0 Step : My : best Put guess a pror of on (,Σ) Step E[ a: ] Put a 0 pror on Σ (ν 0 -m-)σ ~ IW(ν 0, (ν 0 -m-)σ 0 ) Step b: Put a pror on Σ Σ ~ N( 0, Σ / κ 0 ) Notce how we are forced to epress our gnorance of proportonally to Σ κ 0 large: I m pretty sure about my guess of 0 Together, Σ and Σ defne a jont dstrbuton on (,Σ) Copyrght 00, 004, Andrew W. Moore Mamum Lelhood: Slde 43 Beng Bayesan: MAP estmates for Gaussans Suppose you have,, ~(..d) N(,Σ) But you don t now or Σ MAP: Whch (,Σ) mamzes p(,σ,, )? Step : Put a pror on (,Σ) Step a: Put a pror on Σ (ν 0 -m-)σ ~ IW(ν 0, (ν 0 -m-)σ 0 ) Step b: Put a pror on Σ Σ ~ N( 0, Σ / κ 0 ) Why do we use ths form of pror? Copyrght 00, 004, Andrew W. Moore Mamum Lelhood: Slde 44
23 Beng Bayesan: MAP estmates for Gaussans Suppose you have,, ~(..d) N(,Σ) But you don t now or Σ MAP: Whch (,Σ) mamzes p(,σ,, )? Step : Put a pror on (,Σ) Step a: Put a pror on Σ (ν 0 -m-)σ ~ IW(ν 0, (ν 0 -m-)σ 0 ) Step b: Put a pror on Σ Σ ~ N( 0, Σ / κ 0 ) Why do we use ths form of pror? Actually, we don t have to But t s computatonally and algebracally convenent t s a conjugate pror. Copyrght 00, 004, Andrew W. Moore Mamum Lelhood: Slde 45 Beng Bayesan: MAP estmates for Gaussans Suppose you have,, ~(..d) N(,Σ) MAP: Whch (,Σ) mamzes p(,σ,, )? Step : Pror: (ν 0 -m-) Σ ~ IW(ν 0, (ν 0 -m-) Σ 0 ), Σ ~ N( 0, Σ / κ 0 ) Step : κ 0 κ ( ν + m ) Σ ( ν 0 + m ) Σ0 + Copyrght 00, 004, Andrew W. Moore Mamum Lelhood: Slde ν ν 0 + κ κ 0 + T 0 ( )( ) + / κ 0 Step 3: Posteror: (ν +m-)σ ~ IW(ν, (ν +m-) Σ ), Σ ~ N(, Σ / κ ) esult: map, E[Σ,, ] Σ ( )( ) 0 + / T 3
24 Beng Bayesan: Loo MAP carefully estmates what these for formulae Gaussans are Suppose you have,, ~(..d) N(,Σ) MAP: Whch (,Σ) statstcs mamzes of the p(,σ data.,, )? Step : Pror: (ν 0 -m-) Σ ~ The IW(ν margnal 0, (ν 0 -m-) dstrbuton Σ 0 ), Σ on ~ N( s a 0, student-t Σ / κ 0 ) One pont of vew: t s pretty academc f > 30 Step : κ κ + ( ν + m ) Σ ( ν 0 + m ) Σ0 + dong. It s all very sensble. Conjugate prors mean pror form and posteror form are same and characterzed by suffcent Copyrght 00, 004, Andrew W. Moore Mamum Lelhood: Slde 47 0 ν ν 0 + κ κ 0 + T 0 ( )( ) + / κ 0 Step 3: Posteror: (ν +m-)σ ~ IW(ν, (ν +m-) Σ ), Σ ~ N(, Σ / κ ) esult: map, E[Σ,, ] Σ ( )( ) 0 + / T Where we re at Categorcal nputs only eal-valued nputs only Med eal / Cat oay Inputs Classfer Predct category Jont BC Naïve BC Dec Tree Inputs Densty Estmator Probablty Jont DE Naïve DE Gauss DE Inputs egressor Predct real no. Copyrght 00, 004, Andrew W. Moore Mamum Lelhood: Slde 48 4
25 What you should now The ecpe for MLE What do we sometmes prefer MLE to MAP? Understand MLE estmaton of Gaussan parameters Understand based estmator versus unbased estmator Apprecate the outlne behnd Bayesan estmaton of Gaussan parameters Copyrght 00, 004, Andrew W. Moore Mamum Lelhood: Slde 49 Useful eercse We d already done some MLE n ths class wthout even tellng you! Suppose categorcal arty-n nputs,, ~(..d.) from a multnomal M(p, p, p n ) where P( j p)p j What s the MLE p(p, p, p n )? Copyrght 00, 004, Andrew W. Moore Mamum Lelhood: Slde 50 5
Instance-Based Learning (a.k.a. memory-based learning) Part I: Nearest Neighbor Classification
Instance-Based earnng (a.k.a. memory-based learnng) Part I: Nearest Neghbor Classfcaton Note to other teachers and users of these sldes. Andrew would be delghted f you found ths source materal useful n
More informationClustering with Gaussian Mixtures
Note to other teachers and users of these sldes. Andrew would be delghted f you found ths source materal useful n gvng your own lectures. Feel free to use these sldes verbatm, or to modfy them to ft your
More informationThe conjugate prior to a Bernoulli is. A) Bernoulli B) Gaussian C) Beta D) none of the above
The conjugate pror to a Bernoull s A) Bernoull B) Gaussan C) Beta D) none of the above The conjugate pror to a Gaussan s A) Bernoull B) Gaussan C) Beta D) none of the above MAP estmates A) argmax θ p(θ
More informationClustering with Gaussian Mixtures
Note to other teachers and users of these sldes. Andrew would be delghted f you found ths source materal useful n gvng your own lectures. Feel free to use these sldes verbatm, or to modfy them to ft your
More informationENG 8801/ Special Topics in Computer Engineering: Pattern Recognition. Memorial University of Newfoundland Pattern Recognition
EG 880/988 - Specal opcs n Computer Engneerng: Pattern Recognton Memoral Unversty of ewfoundland Pattern Recognton Lecture 7 May 3, 006 http://wwwengrmunca/~charlesr Offce Hours: uesdays hursdays 8:30-9:30
More informationCS 2750 Machine Learning. Lecture 5. Density estimation. CS 2750 Machine Learning. Announcements
CS 750 Machne Learnng Lecture 5 Densty estmaton Mlos Hauskrecht mlos@cs.ptt.edu 539 Sennott Square CS 750 Machne Learnng Announcements Homework Due on Wednesday before the class Reports: hand n before
More informationGaussian Mixture Models
Lab Gaussan Mxture Models Lab Objectve: Understand the formulaton of Gaussan Mxture Models (GMMs) and how to estmate GMM parameters. You ve already seen GMMs as the observaton dstrbuton n certan contnuous
More informationExpectation Maximization Mixture Models HMMs
-755 Machne Learnng for Sgnal Processng Mture Models HMMs Class 9. 2 Sep 200 Learnng Dstrbutons for Data Problem: Gven a collecton of eamples from some data, estmate ts dstrbuton Basc deas of Mamum Lelhood
More informationMLE and Bayesian Estimation. Jie Tang Department of Computer Science & Technology Tsinghua University 2012
MLE and Bayesan Estmaton Je Tang Department of Computer Scence & Technology Tsnghua Unversty 01 1 Lnear Regresson? As the frst step, we need to decde how we re gong to represent the functon f. One example:
More informationUsing T.O.M to Estimate Parameter of distributions that have not Single Exponential Family
IOSR Journal of Mathematcs IOSR-JM) ISSN: 2278-5728. Volume 3, Issue 3 Sep-Oct. 202), PP 44-48 www.osrjournals.org Usng T.O.M to Estmate Parameter of dstrbutons that have not Sngle Exponental Famly Jubran
More informationStatistical analysis using matlab. HY 439 Presented by: George Fortetsanakis
Statstcal analyss usng matlab HY 439 Presented by: George Fortetsanaks Roadmap Probablty dstrbutons Statstcal estmaton Fttng data to probablty dstrbutons Contnuous dstrbutons Contnuous random varable X
More informationxp(x µ) = 0 p(x = 0 µ) + 1 p(x = 1 µ) = µ
CSE 455/555 Sprng 2013 Homework 7: Parametrc Technques Jason J. Corso Computer Scence and Engneerng SUY at Buffalo jcorso@buffalo.edu Solutons by Yngbo Zhou Ths assgnment does not need to be submtted and
More informationHowever, since P is a symmetric idempotent matrix, of P are either 0 or 1 [Eigen-values
Fall 007 Soluton to Mdterm Examnaton STAT 7 Dr. Goel. [0 ponts] For the general lnear model = X + ε, wth uncorrelated errors havng mean zero and varance σ, suppose that the desgn matrx X s not necessarly
More informationGenerative classification models
CS 675 Intro to Machne Learnng Lecture Generatve classfcaton models Mlos Hauskrecht mlos@cs.ptt.edu 539 Sennott Square Data: D { d, d,.., dn} d, Classfcaton represents a dscrete class value Goal: learn
More informationMIMA Group. Chapter 2 Bayesian Decision Theory. School of Computer Science and Technology, Shandong University. Xin-Shun SDU
Group M D L M Chapter Bayesan Decson heory Xn-Shun Xu @ SDU School of Computer Scence and echnology, Shandong Unversty Bayesan Decson heory Bayesan decson theory s a statstcal approach to data mnng/pattern
More information1 GSW Iterative Techniques for y = Ax
1 for y = A I m gong to cheat here. here are a lot of teratve technques that can be used to solve the general case of a set of smultaneous equatons (wrtten n the matr form as y = A), but ths chapter sn
More information10-701/ Machine Learning, Fall 2005 Homework 3
10-701/15-781 Machne Learnng, Fall 2005 Homework 3 Out: 10/20/05 Due: begnnng of the class 11/01/05 Instructons Contact questons-10701@autonlaborg for queston Problem 1 Regresson and Cross-valdaton [40
More informationStat 543 Exam 2 Spring 2016
Stat 543 Exam 2 Sprng 206 I have nether gven nor receved unauthorzed assstance on ths exam. Name Sgned Date Name Prnted Ths Exam conssts of questons. Do at least 0 of the parts of the man exam. I wll score
More informationHidden Markov Models
Note to other teachers and users of these sldes. Andrew would be delghted f you found ths source materal useful n gvng your own lectures. Feel free to use these sldes verbatm, or to modfy them to ft your
More informationStat 543 Exam 2 Spring 2016
Stat 543 Exam 2 Sprng 2016 I have nether gven nor receved unauthorzed assstance on ths exam. Name Sgned Date Name Prnted Ths Exam conssts of 11 questons. Do at least 10 of the 11 parts of the man exam.
More informationClassification learning II
Lecture 8 Classfcaton learnng II Mlos Hauskrecht mlos@cs.ptt.edu 539 Sennott Square Logstc regresson model Defnes a lnear decson boundar Dscrmnant functons: g g g g here g z / e z f, g g - s a logstc functon
More informationEcon107 Applied Econometrics Topic 3: Classical Model (Studenmund, Chapter 4)
I. Classcal Assumptons Econ7 Appled Econometrcs Topc 3: Classcal Model (Studenmund, Chapter 4) We have defned OLS and studed some algebrac propertes of OLS. In ths topc we wll study statstcal propertes
More informationMACHINE APPLIED MACHINE LEARNING LEARNING. Gaussian Mixture Regression
11 MACHINE APPLIED MACHINE LEARNING LEARNING MACHINE LEARNING Gaussan Mture Regresson 22 MACHINE APPLIED MACHINE LEARNING LEARNING Bref summary of last week s lecture 33 MACHINE APPLIED MACHINE LEARNING
More informationCIS526: Machine Learning Lecture 3 (Sept 16, 2003) Linear Regression. Preparation help: Xiaoying Huang. x 1 θ 1 output... θ M x M
CIS56: achne Learnng Lecture 3 (Sept 6, 003) Preparaton help: Xaoyng Huang Lnear Regresson Lnear regresson can be represented by a functonal form: f(; θ) = θ 0 0 +θ + + θ = θ = 0 ote: 0 s a dummy attrbute
More informationFirst Year Examination Department of Statistics, University of Florida
Frst Year Examnaton Department of Statstcs, Unversty of Florda May 7, 010, 8:00 am - 1:00 noon Instructons: 1. You have four hours to answer questons n ths examnaton.. You must show your work to receve
More informationThe Gaussian classifier. Nuno Vasconcelos ECE Department, UCSD
he Gaussan classfer Nuno Vasconcelos ECE Department, UCSD Bayesan decson theory recall that we have state of the world X observatons g decson functon L[g,y] loss of predctng y wth g Bayes decson rule s
More information1. Inference on Regression Parameters a. Finding Mean, s.d and covariance amongst estimates. 2. Confidence Intervals and Working Hotelling Bands
Content. Inference on Regresson Parameters a. Fndng Mean, s.d and covarance amongst estmates.. Confdence Intervals and Workng Hotellng Bands 3. Cochran s Theorem 4. General Lnear Testng 5. Measures of
More informationBayesian Learning. Smart Home Health Analytics Spring Nirmalya Roy Department of Information Systems University of Maryland Baltimore County
Smart Home Health Analytcs Sprng 2018 Bayesan Learnng Nrmalya Roy Department of Informaton Systems Unversty of Maryland Baltmore ounty www.umbc.edu Bayesan Learnng ombnes pror knowledge wth evdence to
More informationDepartment of Computer Science Artificial Intelligence Research Laboratory. Iowa State University MACHINE LEARNING
MACHINE LEANING Vasant Honavar Bonformatcs and Computatonal Bology rogram Center for Computatonal Intellgence, Learnng, & Dscovery Iowa State Unversty honavar@cs.astate.edu www.cs.astate.edu/~honavar/
More informationSingle- Parameter Linear Regression
Predctng Real-valued outputs an ntroducton to Regresson Note to other teachers and users of these sldes. Andrew would be delghted f ou found ths source materal useful n gvng our own lectures. Feel free
More informationComputation of Higher Order Moments from Two Multinomial Overdispersion Likelihood Models
Computaton of Hgher Order Moments from Two Multnomal Overdsperson Lkelhood Models BY J. T. NEWCOMER, N. K. NEERCHAL Department of Mathematcs and Statstcs, Unversty of Maryland, Baltmore County, Baltmore,
More informationCS 3710: Visual Recognition Classification and Detection. Adriana Kovashka Department of Computer Science January 13, 2015
CS 3710: Vsual Recognton Classfcaton and Detecton Adrana Kovashka Department of Computer Scence January 13, 2015 Plan for Today Vsual recognton bascs part 2: Classfcaton and detecton Adrana s research
More informationClassification as a Regression Problem
Target varable y C C, C,, ; Classfcaton as a Regresson Problem { }, 3 L C K To treat classfcaton as a regresson problem we should transform the target y nto numercal values; The choce of numercal class
More informationAS-Level Maths: Statistics 1 for Edexcel
1 of 6 AS-Level Maths: Statstcs 1 for Edecel S1. Calculatng means and standard devatons Ths con ndcates the slde contans actvtes created n Flash. These actvtes are not edtable. For more detaled nstructons,
More informationFall 2012 Analysis of Experimental Measurements B. Eisenstein/rev. S. Errede
Fall 0 Analyss of Expermental easurements B. Esensten/rev. S. Errede We now reformulate the lnear Least Squares ethod n more general terms, sutable for (eventually extendng to the non-lnear case, and also
More informationSupport Vector Machines
CS 2750: Machne Learnng Support Vector Machnes Prof. Adrana Kovashka Unversty of Pttsburgh February 17, 2016 Announcement Homework 2 deadlne s now 2/29 We ll have covered everythng you need today or at
More informationComposite Hypotheses testing
Composte ypotheses testng In many hypothess testng problems there are many possble dstrbutons that can occur under each of the hypotheses. The output of the source s a set of parameters (ponts n a parameter
More information2E Pattern Recognition Solutions to Introduction to Pattern Recognition, Chapter 2: Bayesian pattern classification
E395 - Pattern Recognton Solutons to Introducton to Pattern Recognton, Chapter : Bayesan pattern classfcaton Preface Ths document s a soluton manual for selected exercses from Introducton to Pattern Recognton
More informationMaximum Likelihood Estimation
Maxmum Lkelhood Estmaton INFO-2301: Quanttatve Reasonng 2 Mchael Paul and Jordan Boyd-Graber MARCH 7, 2017 INFO-2301: Quanttatve Reasonng 2 Paul and Boyd-Graber Maxmum Lkelhood Estmaton 1 of 9 Why MLE?
More informationLinear Feature Engineering 11
Lnear Feature Engneerng 11 2 Least-Squares 2.1 Smple least-squares Consder the followng dataset. We have a bunch of nputs x and correspondng outputs y. The partcular values n ths dataset are x y 0.23 0.19
More information4 Analysis of Variance (ANOVA) 5 ANOVA. 5.1 Introduction. 5.2 Fixed Effects ANOVA
4 Analyss of Varance (ANOVA) 5 ANOVA 51 Introducton ANOVA ANOVA s a way to estmate and test the means of multple populatons We wll start wth one-way ANOVA If the populatons ncluded n the study are selected
More informationChapter 5. Solution of System of Linear Equations. Module No. 6. Solution of Inconsistent and Ill Conditioned Systems
Numercal Analyss by Dr. Anta Pal Assstant Professor Department of Mathematcs Natonal Insttute of Technology Durgapur Durgapur-713209 emal: anta.bue@gmal.com 1 . Chapter 5 Soluton of System of Lnear Equatons
More informationHomework Assignment 3 Due in class, Thursday October 15
Homework Assgnment 3 Due n class, Thursday October 15 SDS 383C Statstcal Modelng I 1 Rdge regresson and Lasso 1. Get the Prostrate cancer data from http://statweb.stanford.edu/~tbs/elemstatlearn/ datasets/prostate.data.
More informationParametric fractional imputation for missing data analysis. Jae Kwang Kim Survey Working Group Seminar March 29, 2010
Parametrc fractonal mputaton for mssng data analyss Jae Kwang Km Survey Workng Group Semnar March 29, 2010 1 Outlne Introducton Proposed method Fractonal mputaton Approxmaton Varance estmaton Multple mputaton
More informationWhich Separator? Spring 1
Whch Separator? 6.034 - Sprng 1 Whch Separator? Mamze the margn to closest ponts 6.034 - Sprng Whch Separator? Mamze the margn to closest ponts 6.034 - Sprng 3 Margn of a pont " # y (w $ + b) proportonal
More informationWhy Bayesian? 3. Bayes and Normal Models. State of nature: class. Decision rule. Rev. Thomas Bayes ( ) Bayes Theorem (yes, the famous one)
Why Bayesan? 3. Bayes and Normal Models Alex M. Martnez alex@ece.osu.edu Handouts Handoutsfor forece ECE874 874Sp Sp007 If all our research (n PR was to dsappear and you could only save one theory, whch
More information3.1 Expectation of Functions of Several Random Variables. )' be a k-dimensional discrete or continuous random vector, with joint PMF p (, E X E X1 E X
Statstcs 1: Probablty Theory II 37 3 EPECTATION OF SEVERAL RANDOM VARIABLES As n Probablty Theory I, the nterest n most stuatons les not on the actual dstrbuton of a random vector, but rather on a number
More informationConjugacy and the Exponential Family
CS281B/Stat241B: Advanced Topcs n Learnng & Decson Makng Conjugacy and the Exponental Famly Lecturer: Mchael I. Jordan Scrbes: Bran Mlch 1 Conjugacy In the prevous lecture, we saw conjugate prors for the
More informationDifferentiating Gaussian Processes
Dfferentatng Gaussan Processes Andrew McHutchon Aprl 17, 013 1 Frst Order Dervatve of the Posteror Mean The posteror mean of a GP s gven by, f = x, X KX, X 1 y x, X α 1 Only the x, X term depends on the
More informationLogistic Regression. CAP 5610: Machine Learning Instructor: Guo-Jun QI
Logstc Regresson CAP 561: achne Learnng Instructor: Guo-Jun QI Bayes Classfer: A Generatve model odel the posteror dstrbuton P(Y X) Estmate class-condtonal dstrbuton P(X Y) for each Y Estmate pror dstrbuton
More informationLogistic Regression Maximum Likelihood Estimation
Harvard-MIT Dvson of Health Scences and Technology HST.951J: Medcal Decson Support, Fall 2005 Instructors: Professor Lucla Ohno-Machado and Professor Staal Vnterbo 6.873/HST.951 Medcal Decson Support Fall
More informationDepartment of Statistics University of Toronto STA305H1S / 1004 HS Design and Analysis of Experiments Term Test - Winter Solution
Department of Statstcs Unversty of Toronto STA35HS / HS Desgn and Analyss of Experments Term Test - Wnter - Soluton February, Last Name: Frst Name: Student Number: Instructons: Tme: hours. Ads: a non-programmable
More informationPrimer on High-Order Moment Estimators
Prmer on Hgh-Order Moment Estmators Ton M. Whted July 2007 The Errors-n-Varables Model We wll start wth the classcal EIV for one msmeasured regressor. The general case s n Erckson and Whted Econometrc
More informationChapter 14 Simple Linear Regression
Chapter 4 Smple Lnear Regresson Chapter 4 - Smple Lnear Regresson Manageral decsons often are based on the relatonshp between two or more varables. Regresson analss can be used to develop an equaton showng
More informationIssues To Consider when Estimating Health Care Costs with Generalized Linear Models (GLMs): To Gamma/Log Or Not To Gamma/Log? That Is The New Question
Issues To Consder when Estmatng Health Care Costs wth Generalzed Lnear Models (GLMs): To Gamma/Log Or Not To Gamma/Log? That Is The New Queston ISPOR 20th Annual Internatonal Meetng May 19, 2015 Jalpa
More informationImage classification. Given the bag-of-features representations of images from different classes, how do we learn a model for distinguishing i them?
Image classfcaton Gven te bag-of-features representatons of mages from dfferent classes ow do we learn a model for dstngusng tem? Classfers Learn a decson rule assgnng bag-offeatures representatons of
More informationComparison of Regression Lines
STATGRAPHICS Rev. 9/13/2013 Comparson of Regresson Lnes Summary... 1 Data Input... 3 Analyss Summary... 4 Plot of Ftted Model... 6 Condtonal Sums of Squares... 6 Analyss Optons... 7 Forecasts... 8 Confdence
More informationLearning from Data 1 Naive Bayes
Learnng from Data 1 Nave Bayes Davd Barber dbarber@anc.ed.ac.uk course page : http://anc.ed.ac.uk/ dbarber/lfd1/lfd1.html c Davd Barber 2001, 2002 1 Learnng from Data 1 : c Davd Barber 2001,2002 2 1 Why
More informationProperties of Least Squares
Week 3 3.1 Smple Lnear Regresson Model 3. Propertes of Least Squares Estmators Y Y β 1 + β X + u weekly famly expendtures X weekly famly ncome For a gven level of x, the expected level of food expendtures
More informationUNIVERSITY OF TORONTO Faculty of Arts and Science. December 2005 Examinations STA437H1F/STA1005HF. Duration - 3 hours
UNIVERSITY OF TORONTO Faculty of Arts and Scence December 005 Examnatons STA47HF/STA005HF Duraton - hours AIDS ALLOWED: (to be suppled by the student) Non-programmable calculator One handwrtten 8.5'' x
More informationMixture o f of Gaussian Gaussian clustering Nov
Mture of Gaussan clusterng Nov 11 2009 Soft vs hard lusterng Kmeans performs Hard clusterng: Data pont s determnstcally assgned to one and only one cluster But n realty clusters may overlap Soft-clusterng:
More informationβ0 + β1xi and want to estimate the unknown
SLR Models Estmaton Those OLS Estmates Estmators (e ante) v. estmates (e post) The Smple Lnear Regresson (SLR) Condtons -4 An Asde: The Populaton Regresson Functon B and B are Lnear Estmators (condtonal
More informationStatistical Foundations of Pattern Recognition
Statstcal Foundatons of Pattern Recognton Learnng Objectves Bayes Theorem Decson-mang Confdence factors Dscrmnants The connecton to neural nets Statstcal Foundatons of Pattern Recognton NDE measurement
More informationDepartment of Quantitative Methods & Information Systems. Time Series and Their Components QMIS 320. Chapter 6
Department of Quanttatve Methods & Informaton Systems Tme Seres and Ther Components QMIS 30 Chapter 6 Fall 00 Dr. Mohammad Zanal These sldes were modfed from ther orgnal source for educatonal purpose only.
More informationOutline. Multivariate Parametric Methods. Multivariate Data. Basic Multivariate Statistics. Steven J Zeil
Outlne Multvarate Parametrc Methods Steven J Zel Old Domnon Unv. Fall 2010 1 Multvarate Data 2 Multvarate ormal Dstrbuton 3 Multvarate Classfcaton Dscrmnants Tunng Complexty Dscrete Features 4 Multvarate
More informationLINEAR REGRESSION ANALYSIS. MODULE IX Lecture Multicollinearity
LINEAR REGRESSION ANALYSIS MODULE IX Lecture - 30 Multcollnearty Dr. Shalabh Department of Mathematcs and Statstcs Indan Insttute of Technology Kanpur 2 Remedes for multcollnearty Varous technques have
More informationLecture 6: Introduction to Linear Regression
Lecture 6: Introducton to Lnear Regresson An Manchakul amancha@jhsph.edu 24 Aprl 27 Lnear regresson: man dea Lnear regresson can be used to study an outcome as a lnear functon of a predctor Example: 6
More informationMATH 829: Introduction to Data Mining and Analysis The EM algorithm (part 2)
1/16 MATH 829: Introducton to Data Mnng and Analyss The EM algorthm (part 2) Domnque Gullot Departments of Mathematcal Scences Unversty of Delaware Aprl 20, 2016 Recall 2/16 We are gven ndependent observatons
More informationRandomness and Computation
Randomness and Computaton or, Randomzed Algorthms Mary Cryan School of Informatcs Unversty of Ednburgh RC 208/9) Lecture 0 slde Balls n Bns m balls, n bns, and balls thrown unformly at random nto bns usually
More informationLecture 12: Classification
Lecture : Classfcaton g Dscrmnant functons g The optmal Bayes classfer g Quadratc classfers g Eucldean and Mahalanobs metrcs g K Nearest Neghbor Classfers Intellgent Sensor Systems Rcardo Guterrez-Osuna
More informationChapter 2 - The Simple Linear Regression Model S =0. e i is a random error. S β2 β. This is a minimization problem. Solution is a calculus exercise.
Chapter - The Smple Lnear Regresson Model The lnear regresson equaton s: where y + = β + β e for =,..., y and are observable varables e s a random error How can an estmaton rule be constructed for the
More informatione i is a random error
Chapter - The Smple Lnear Regresson Model The lnear regresson equaton s: where + β + β e for,..., and are observable varables e s a random error How can an estmaton rule be constructed for the unknown
More informationBasically, if you have a dummy dependent variable you will be estimating a probability.
ECON 497: Lecture Notes 13 Page 1 of 1 Metropoltan State Unversty ECON 497: Research and Forecastng Lecture Notes 13 Dummy Dependent Varable Technques Studenmund Chapter 13 Bascally, f you have a dummy
More informationAn Experiment/Some Intuition (Fall 2006): Lecture 18 The EM Algorithm heads coin 1 tails coin 2 Overview Maximum Likelihood Estimation
An Experment/Some Intuton I have three cons n my pocket, 6.864 (Fall 2006): Lecture 18 The EM Algorthm Con 0 has probablty λ of heads; Con 1 has probablty p 1 of heads; Con 2 has probablty p 2 of heads
More informationLinear regression. Regression Models. Chapter 11 Student Lecture Notes Regression Analysis is the
Chapter 11 Student Lecture Notes 11-1 Lnear regresson Wenl lu Dept. Health statstcs School of publc health Tanjn medcal unversty 1 Regresson Models 1. Answer What Is the Relatonshp Between the Varables?.
More informationj) = 1 (note sigma notation) ii. Continuous random variable (e.g. Normal distribution) 1. density function: f ( x) 0 and f ( x) dx = 1
Random varables Measure of central tendences and varablty (means and varances) Jont densty functons and ndependence Measures of assocaton (covarance and correlaton) Interestng result Condtonal dstrbutons
More information+, where 0 x N - n. k k
CO 745, Mdterm Len Cabrera. A multle choce eam has questons, each of whch has ossble answers. A student nows the correct answer to n of these questons. For the remanng - n questons, he checs the answers
More informationPredictive Analytics : QM901.1x Prof U Dinesh Kumar, IIMB. All Rights Reserved, Indian Institute of Management Bangalore
Sesson Outlne Introducton to classfcaton problems and dscrete choce models. Introducton to Logstcs Regresson. Logstc functon and Logt functon. Maxmum Lkelhood Estmator (MLE) for estmaton of LR parameters.
More informationCHAPTER 3: BAYESIAN DECISION THEORY
HATER 3: BAYESIAN DEISION THEORY Decson mang under uncertanty 3 Data comes from a process that s completely not nown The lac of nowledge can be compensated by modelng t as a random process May be the underlyng
More informationIntegrals and Invariants of Euler-Lagrange Equations
Lecture 16 Integrals and Invarants of Euler-Lagrange Equatons ME 256 at the Indan Insttute of Scence, Bengaluru Varatonal Methods and Structural Optmzaton G. K. Ananthasuresh Professor, Mechancal Engneerng,
More informationProbability Density Function Estimation by different Methods
EEE 739Q SPRIG 00 COURSE ASSIGMET REPORT Probablty Densty Functon Estmaton by dfferent Methods Vas Chandraant Rayar Abstract The am of the assgnment was to estmate the probablty densty functon (PDF of
More informationSDMML HT MSc Problem Sheet 4
SDMML HT 06 - MSc Problem Sheet 4. The recever operatng characterstc ROC curve plots the senstvty aganst the specfcty of a bnary classfer as the threshold for dscrmnaton s vared. Let the data space be
More informationLogistic Classifier CISC 5800 Professor Daniel Leeds
lon 9/7/8 Logstc Classfer CISC 58 Professor Danel Leeds Classfcaton strategy: generatve vs. dscrmnatve Generatve, e.g., Bayes/Naïve Bayes: 5 5 Identfy probablty dstrbuton for each class Determne class
More informationLecture 9: Linear regression: centering, hypothesis testing, multiple covariates, and confounding
Recall: man dea of lnear regresson Lecture 9: Lnear regresson: centerng, hypothess testng, multple covarates, and confoundng Sandy Eckel seckel@jhsph.edu 6 May 8 Lnear regresson can be used to study an
More informationLecture 9: Linear regression: centering, hypothesis testing, multiple covariates, and confounding
Lecture 9: Lnear regresson: centerng, hypothess testng, multple covarates, and confoundng Sandy Eckel seckel@jhsph.edu 6 May 008 Recall: man dea of lnear regresson Lnear regresson can be used to study
More informationBayesian predictive Configural Frequency Analysis
Psychologcal Test and Assessment Modelng, Volume 54, 2012 (3), 285-292 Bayesan predctve Confgural Frequency Analyss Eduardo Gutérrez-Peña 1 Abstract Confgural Frequency Analyss s a method for cell-wse
More informationMachine learning: Density estimation
CS 70 Foundatons of AI Lecture 3 Machne learnng: ensty estmaton Mlos Hauskrecht mlos@cs.ptt.edu 539 Sennott Square ata: ensty estmaton {.. n} x a vector of attrbute values Objectve: estmate the model of
More informationStatistics Chapter 4
Statstcs Chapter 4 "There are three knds of les: les, damned les, and statstcs." Benjamn Dsrael, 1895 (Brtsh statesman) Gaussan Dstrbuton, 4-1 If a measurement s repeated many tmes a statstcal treatment
More informationx i1 =1 for all i (the constant ).
Chapter 5 The Multple Regresson Model Consder an economc model where the dependent varable s a functon of K explanatory varables. The economc model has the form: y = f ( x,x,..., ) xk Approxmate ths by
More informationHidden Markov Models & The Multivariate Gaussian (10/26/04)
CS281A/Stat241A: Statstcal Learnng Theory Hdden Markov Models & The Multvarate Gaussan (10/26/04) Lecturer: Mchael I. Jordan Scrbes: Jonathan W. Hu 1 Hdden Markov Models As a bref revew, hdden Markov models
More informationGrover s Algorithm + Quantum Zeno Effect + Vaidman
Grover s Algorthm + Quantum Zeno Effect + Vadman CS 294-2 Bomb 10/12/04 Fall 2004 Lecture 11 Grover s algorthm Recall that Grover s algorthm for searchng over a space of sze wors as follows: consder the
More informationStat 642, Lecture notes for 01/27/ d i = 1 t. n i t nj. n j
Stat 642, Lecture notes for 01/27/05 18 Rate Standardzaton Contnued: Note that f T n t where T s the cumulatve follow-up tme and n s the number of subjects at rsk at the mdpont or nterval, and d s the
More informationStatistics for Business and Economics
Statstcs for Busness and Economcs Chapter 11 Smple Regresson Copyrght 010 Pearson Educaton, Inc. Publshng as Prentce Hall Ch. 11-1 11.1 Overvew of Lnear Models n An equaton can be ft to show the best lnear
More informationStatistics MINITAB - Lab 2
Statstcs 20080 MINITAB - Lab 2 1. Smple Lnear Regresson In smple lnear regresson we attempt to model a lnear relatonshp between two varables wth a straght lne and make statstcal nferences concernng that
More informationThe EM Algorithm (Dempster, Laird, Rubin 1977) The missing data or incomplete data setting: ODL(φ;Y ) = [Y;φ] = [Y X,φ][X φ] = X
The EM Algorthm (Dempster, Lard, Rubn 1977 The mssng data or ncomplete data settng: An Observed Data Lkelhood (ODL that s a mxture or ntegral of Complete Data Lkelhoods (CDL. (1a ODL(;Y = [Y;] = [Y,][
More informationPattern Classification
Pattern Classfcaton All materals n these sldes ere taken from Pattern Classfcaton (nd ed) by R. O. Duda, P. E. Hart and D. G. Stork, John Wley & Sons, 000 th the permsson of the authors and the publsher
More informationLecture 10 Support Vector Machines II
Lecture 10 Support Vector Machnes II 22 February 2016 Taylor B. Arnold Yale Statstcs STAT 365/665 1/28 Notes: Problem 3 s posted and due ths upcomng Frday There was an early bug n the fake-test data; fxed
More informationTHE ROYAL STATISTICAL SOCIETY 2006 EXAMINATIONS SOLUTIONS HIGHER CERTIFICATE
THE ROYAL STATISTICAL SOCIETY 6 EXAMINATIONS SOLUTIONS HIGHER CERTIFICATE PAPER I STATISTICAL THEORY The Socety provdes these solutons to assst canddates preparng for the eamnatons n future years and for
More informationLimited Dependent Variables
Lmted Dependent Varables. What f the left-hand sde varable s not a contnuous thng spread from mnus nfnty to plus nfnty? That s, gven a model = f (, β, ε, where a. s bounded below at zero, such as wages
More informationSpace of ML Problems. CSE 473: Artificial Intelligence. Parameter Estimation and Bayesian Networks. Learning Topics
/7/7 CSE 73: Artfcal Intellgence Bayesan - Learnng Deter Fox Sldes adapted from Dan Weld, Jack Breese, Dan Klen, Daphne Koller, Stuart Russell, Andrew Moore & Luke Zettlemoyer What s Beng Learned? Space
More information