Lecture 8: Multiclass Classification (I)

Size: px
Start display at page:

Download "Lecture 8: Multiclass Classification (I)"

Transcription

1 Bayes Rule fr Multiclass Prblems Traditinal Methds fr Multiclass Prblems Linear Regressin Mdels Lecture 8: Multiclass Classificatin (I) Ha Helen Zhang Fall 07 Ha Helen Zhang Lecture 8: Multiclass Classificatin (I)

2 Bayes Rule fr Multiclass Prblems Traditinal Methds fr Multiclass Prblems Linear Regressin Mdels Lecture 8: Multiclass Classificatin (I) Ha Helen Zhang Fall 07 Ha Helen Zhang Lecture 8: Multiclass Classificatin (I)

3 Bayes Rule fr Multiclass Prblems Traditinal Methds fr Multiclass Prblems Linear Regressin Mdels Multiclass Classificatin General setup Bayes rule fr multiclass prblems under equal csts under unequal csts Linear multiclass methds Linear regressin mdel Liner discriminant analysis (LDA) Multiple lgistic regressin Mdels Ha Helen Zhang Lecture 8: Multiclass Classificatin (I)

4 Bayes Rule fr Multiclass Prblems Traditinal Methds fr Multiclass Prblems Linear Regressin Mdels Multiclass Prblems Class label Y {,..., K}, K. The classifier f : R d {,..., K}. Examples: zip cde classificatin, multiple-type cancer classificatin, vwel recgnitin The lss functin C(k, l) = cst f classifying a sample in class k t class l. In general, C(k, k) = 0 fr any k =,, K. The lss can be described as an K K matrix. Fr example, K =, 0 C = 0. 0 Ha Helen Zhang Lecture 8: Multiclass Classificatin (I)

5 Bayes Rule fr Multiclass Prblems Traditinal Methds fr Multiclass Prblems Linear Regressin Mdels Terminlgies Prir class prbabilities (marginal distributin f Y ): π k = P(Y = k), k =,, K Cnditinal distributin f X given Y : assume that g k is the cnditinal density f X given Y = k, k =,, K. Marginal density f X (a mixture): g(x) = K π k g k (x), k =,, K. k= Psterir class prbabilities (cnditinal dist. f Y given X): P(Y = k X = x) = π kg k (x), k =,, K. g(x) Ha Helen Zhang Lecture 8: Multiclass Classificatin (I)

6 Bayes Rule General Setup Bayes Rule fr Multiclass Prblems Traditinal Methds fr Multiclass Prblems Linear Regressin Mdels The ptimal (Bayes) rule aims t minimize the average lss functin ver the ppulatin where φ B (x) = arg min E X,Y C(Y, f (X)) f = arg min E X E Y X C(Y, f (X)), f E Y X L(Y, f (X)) = K K I (f (X) = k) C(l, k)p(y = l X). k= l= Ha Helen Zhang Lecture 8: Multiclass Classificatin (I)

7 Bayes Rule fr Multiclass Prblems Traditinal Methds fr Multiclass Prblems Linear Regressin Mdels Bayes Rule Under Equal Csts If C(k, l) = I (k l), then R[f ] = E X E Y X I (Y f (X )) and E Y X C(Y, f (X)) = = K K I (f (X) = k) P(Y = l X ) k= l k K I (f (X) = k)[ P(Y = k X)] l=k Given x, the minimizer f E Y X=x C(Y, f (X)) is f (x) = k, where The Bayes rule is k = arg min k=,..,k P(Y = k x). φ B (x) = k if P(Y = k x) = max P(Y = k x), k=,...,k which assigns x t the mst prbable class using Pr(Y = k X = x). Ha Helen Zhang Lecture 8: Multiclass Classificatin (I)

8 Bayes Rule fr Multiclass Prblems Traditinal Methds fr Multiclass Prblems Linear Regressin Mdels Bayes Rule Under Unequal Csts Fr the general lss functin, the Bayes rule can be derived as φ B (x) = k if k = arg min k=,...,k K C(l, k)p(y = l x). l= There is nt a simple analytic frm fr the slutin. Ha Helen Zhang Lecture 8: Multiclass Classificatin (I)

9 Bayes Rule fr Multiclass Prblems Traditinal Methds fr Multiclass Prblems Linear Regressin Mdels Decisin (Discriminating) Functins Fr multiclass prblems, we generally need t estimate multiple discriminant functins f k (x), k =,, K Each f k (x) is assciated with class k. f k (x) represents the evidence strength f a sample (x, y) belnging t class k. The decisin rule cnstructed using f k s is ˆf (x) = k, where k = arg max k=,...,k f k(x). The decisin bundary f the classificatin rule ˆf between class k and class l is defined as {x : ˆf k (x) = ˆf l (x)}. Ha Helen Zhang Lecture 8: Multiclass Classificatin (I)

10 Bayes Rule fr Multiclass Prblems Traditinal Methds fr Multiclass Prblems Linear Regressin Mdels Traditinal Methds: Divide and Cnquer Main ideas: (i) Decmpse the multiclass classificatin prblem int multiple binary classificatin prblems. (ii) Use the majrity vting principle (a cmbined decisin frm the cmmittee) t predict the label Cmmn appraches: simple but effective One-vs-rest (ne-vs-all) appraches Pairwise (ne-vs-ne, all-vs-all) appraches Ha Helen Zhang Lecture 8: Multiclass Classificatin (I)

11 Bayes Rule fr Multiclass Prblems Traditinal Methds fr Multiclass Prblems Linear Regressin Mdels One-vs-rest Apprach One f the simplest multiclass classifier; cmmnly used in SVMs; als knwn as the ne-vs-all (OVA) apprach (i) Slve K different binary prblems: classify class k versus the rest classes fr k =,, K. (ii) Assign a test sample t the class giving the largest f k (x) (mst psitive) value, where f k (x) is the slutin frm the kth prblem Prperties: Very simple t implement, perfrm well in practice Nt ptimal (asympttically): the decisin rule is nt Fisher cnsistent if there is n dminating class (i.e. arg max p k (x) < ). Read: Rifkin and Klautau (004) In Defense f One-vs-all Classificatin Ha Helen Zhang Lecture 8: Multiclass Classificatin (I)

12 Bayes Rule fr Multiclass Prblems Traditinal Methds fr Multiclass Prblems Linear Regressin Mdels Pairwise Apprach Als knwn as all-vs-all (AVA) apprach (i) Slve ( K ) different binary prblems: classify class k versus class j fr all j k. Each classifier is called g ij. (ii) Fr predictin at a pint, each classifier is queried nce and issues a vte. The class with the maximum number f (weighted) vtes is the winner. Prperties: Training prcess is efficient, by dealing with small binary prblems. If K is big, there are t many prblems t slve. If K = 0, we need t train 45 binary classifiers. Simple t implement; perfrm cmpetitively in practice. Read: Park and Furnkranz (007) Efficient Pairwise Classificatin Ha Helen Zhang Lecture 8: Multiclass Classificatin (I)

13 Bayes Rule fr Multiclass Prblems Traditinal Methds fr Multiclass Prblems Linear Regressin Mdels Multivariate Linear Regressin Multiple Lgistic Regressin Linear Classifier fr Multiclass Prblems Assume the decisin bundaries are linear. Cmmn linear classificatin methds: Multivariate linear regressin methds Linear lg-dds (lgit) mdels Linear discriminant analysis (LDA) Multiple lgistic regressin Separating hyperplane explicitly mdel the bundaries as linear Perceptrn mdel, SVMs Ha Helen Zhang Lecture 8: Multiclass Classificatin (I)

14 Bayes Rule fr Multiclass Prblems Traditinal Methds fr Multiclass Prblems Linear Regressin Mdels Cding fr Linear Regressin Multivariate Linear Regressin Multiple Lgistic Regressin Fr each class k, cde it using an indicatr variable Z k : Z k = if Y = k; Z k = 0 if Y k. The respnse y is cded using a vectr z = (z,..., z K ). If Y = k, nly the kth cmpnent f z =. The entire training samples can be cded using a n K matrix, dented by Z. Call the matrix Z the indicatr respnse matrix. Ha Helen Zhang Lecture 8: Multiclass Classificatin (I)

15 Bayes Rule fr Multiclass Prblems Traditinal Methds fr Multiclass Prblems Linear Regressin Mdels Multivariate Lineaer Regressin Multivariate Linear Regressin Multiple Lgistic Regressin Cnsider the multivariate mdel Z = XB + E, r Z k = f k Xb k + ɛ k, k =,, K. Z k is the kth clumn f the indicatr respnse matrix Z. X is n (d + ) input matrix E is n K matrix f errrs. B is (d + ) K matrix f parameters, b k is the kth clumn f B. Ha Helen Zhang Lecture 8: Multiclass Classificatin (I)

16 Bayes Rule fr Multiclass Prblems Traditinal Methds fr Multiclass Prblems Linear Regressin Mdels Ordinary Least Square Estimates Multivariate Linear Regressin Multiple Lgistic Regressin Minimize the residual sum f squares K n RSS(B) = [z ik β 0k k= i= d β jk x ij ] j= = tr[(z XB) T (Z XB)]. The LS estimates has the same frm as univariate case ˆB = (X T X ) X T Z, Ẑ = X (X T X ) X T Z X ˆB. Cefficient b k fr y k is same as univariate OLS estimates. (Multiple utputs d nt affect the OLS estimates). [ T Fr x, cmpute ˆf (x) = ( x) ˆB] and classify it as ˆf (x) = argmax k=,...,k ˆfk (x). Ha Helen Zhang Lecture 8: Multiclass Classificatin (I)

17 Bayes Rule fr Multiclass Prblems Traditinal Methds fr Multiclass Prblems Linear Regressin Mdels Abut Multiple Linear Regressin Multivariate Linear Regressin Multiple Lgistic Regressin Justificatin: Each ˆf k is an estimate f cnditinal expectatin f Y k, which is the cnditinal prbability f (X, Y ) belnging t the class k E(Y k X = x) = Pr(Y = k X = x). If the intercept is in the mdel (clumn f in X), then K k= ˆf k (x) = fr any x. If the linear assumptin is apprpriate, then ˆf k give a cnsistent estimate f the prbability fr k =,, K, as the sample size n ges t infinity. Ha Helen Zhang Lecture 8: Multiclass Classificatin (I)

18 Bayes Rule fr Multiclass Prblems Traditinal Methds fr Multiclass Prblems Linear Regressin Mdels Limitatins f Linear Regressin Multivariate Linear Regressin Multiple Lgistic Regressin There is n guarantee that all ˆf k [0, ]; sme can be negative r greater than. Linear structure f X s can be rigid. Serius masking prblem fr K due t rigid linear structure. Ha Helen Zhang Lecture 8: Multiclass Classificatin (I)

19 Bayes Rule fr Multiclass Prblems Traditinal Methds fr Multiclass Prblems Linear Regressin Mdels Multivariate Linear Regressin Multiple Lgistic Regressin Ð Ñ ÒØ Ó ËØ Ø Ø Ð Ä ÖÒ Ò À Ø Ì Ö Ò ² Ö Ñ Ò ¾¼¼½ ÔØ Ö Linear Regressin Ö ÔÐ Ñ ÒØ ½ ½ ¾ ¾ ÙÖ º¾ Ì Ø ÓÑ ÖÓÑ Ø Ö Ð Ò ÁÊ ¾ Ò Ö ÐÝ Ô Ö Ø Ý Ð Ò Ö ÓÒ ÓÙÒ ¹ Ö º Ì Ö Ø ÔÐÓØ ÓÛ Ø ÓÙÒ Ö ÓÙÒ Ý Ð Ò Ö Ö Ñ Ò ÒØ Ò ÐÝ º Ì Ð Ø ÔÐÓØ ÓÛ Ø ÓÙÒ Ö ÓÙÒ Ý Ð Ò Ö Ö Ö ÓÒ Ó Ø Ò ¹ ØÓÖ Ö ÔÓÒ Ú Ö Ð º Ì Ñ Ð Ð ÓÑÔÐ Ø ÐÝ Ñ Ò Ú Ö ÓÑ Ò Ø µº Ha Helen Zhang Lecture 8: Multiclass Classificatin (I)

20 General Setup Bayes Rule fr Multiclass Prblems Traditinal Methds fr Multiclass Prblems Linear Regressin Mdels Multivariate Linear Regressin Multiple Lgistic Regressin Ð Ñ ÒØ Ó ËØ Ø Ø Ð Ä ÖÒ Ò À Ø Ì Ö Ò ² Ö Ñ Ò ¾¼¼½ ÔØ Ö Degree = ; Errr = 0.5 Degree = ; Errr = ÙÖ º Ì «Ø Ó Ñ Ò ÓÒ Ð Ò Ö Ö Ö ¹ ÓÒ Ò ÁÊ ÓÖ Ø Ö ¹Ð ÔÖÓ Ð Ñº Ì ÖÙ ÔÐÓØ Ø Ø Ò Ø Ø ÔÓ Ø ÓÒ Ò Ð Ñ Ñ Ö Ô Ó Ó ÖÚ Ø ÓÒº Ì Ø Ö ÙÖÚ Ò Ô Ò Ð Ö Ø ØØ Ö Ö ÓÒ ØÓ Ø Ø Ö ¹Ð Ò ØÓÖ Ú Ö ¹ Ð ÓÖ Ü ÑÔÐ ÓÖ Ø Ö Ð Ý Ö ½ ÓÖ Ø Ö Ó ÖÚ Ø ÓÒ Ò ¼ ÓÖ Ø Ö Ò Ò ÐÙ º Ì Ø Ö Ð Ò Ö Ò ÕÙ Ö Ø ÔÓÐÝÒÓÑ Ð º ÓÚ ÔÐÓØ Ø ØÖ Ò Ò ÖÖÓÖ Ö Ø º Ì Ý ÖÖÓÖ Ö Ø ¼ ¼¾ ÓÖ Ø ÔÖÓ Ð Ñ Ø Ä ÖÖÓÖ Ö Ø º Ha Helen Zhang Lecture 8: Multiclass Classificatin (I)

21 Bayes Rule fr Multiclass Prblems Traditinal Methds fr Multiclass Prblems Linear Regressin Mdels Abut Masking Prblems Multivariate Linear Regressin Multiple Lgistic Regressin One-dimensinal example: (see the previus figure) Prjected the data nt the line jining the three centrid There is n infrmatin in the nrthwest-sutheast directin Three curves are regressin lines The left panel is fr the linear regressin fit: the middle class is never dminant! The right panel is fr the quadratic fit (masking prblem is slved!) Ha Helen Zhang Lecture 8: Multiclass Classificatin (I)

22 Bayes Rule fr Multiclass Prblems Traditinal Methds fr Multiclass Prblems Linear Regressin Mdels Hw t Slve Masking Prblems Multivariate Linear Regressin Multiple Lgistic Regressin In the previus example with K = Using Quadratic regressin rather than linear regressin Linear regressin errr 0.5; Quadratic regressin errr 0.0; Bayes errr In general, If K classes are lined up, we need plynmial terms up t degree K t reslve the masking prblem. In the wrst case, need O(p K ) terms. Thugh LDA is als based n linear functins, it des nt suffer frm masking prblems. Ha Helen Zhang Lecture 8: Multiclass Classificatin (I)

23 Bayes Rule fr Multiclass Prblems Traditinal Methds fr Multiclass Prblems Linear Regressin Mdels Large K, Small d Prblem Multivariate Linear Regressin Multiple Lgistic Regressin Masking ften ccurs fr large K and small d. Example: Vwel data K = and d = 0. (Tw-dimensinal fr training data using LDA prjectin) A difficult classificatin prblem, as the class verlap is cnsiderate the best methds achieve 40% test errr Methd Traing Errr Test Errr Linear regressin LDA QDA Lgistic regressin Ha Helen Zhang Lecture 8: Multiclass Classificatin (I)

24 General Setup Bayes Rule fr Multiclass Prblems Traditinal Methds fr Multiclass Prblems Linear Regressin Mdels Multivariate Linear Regressin Multiple Lgistic Regressin Elements f Statistical Learning cflhastie, Tibshirani & Friedman 00 Chapter 4 Crdinate fr Training Data Crdinate fr Training Data Figure 4.4: A tw-dimensinal plt f the vwel training data. There are eleven classes with X IR 0, and this is the best view in terms f a LDA mdel (Sectin 4..). The heavy circles are the prjected mean vectrs fr each class. The class verlap is cnsiderable. Ha Helen Zhang Lecture 8: Multiclass Classificatin (I)

25 Bayes Rule fr Multiclass Prblems Traditinal Methds fr Multiclass Prblems Linear Regressin Mdels Multivariate Linear Regressin Multiple Lgistic Regressin Mdel Assumptins fr Mdel Setup Let π k be prir prbabilities f class k Let g k (x) be the class-cnditinal densities f X in class k. The psterir prbability Pr(Y = k X = x) = g k(x)π k K l= g l(x)π l. Mdel Assumptins: Assume each class density is multivariate Gaussian N(µ k, Σ k ). Further, we assume Σ k = Σ fr all k (equal cvariance required!) Ha Helen Zhang Lecture 8: Multiclass Classificatin (I)

26 Bayes Rule fr Multiclass Prblems Traditinal Methds fr Multiclass Prblems Linear Regressin Mdels Linear Discriminant Functin Multivariate Linear Regressin Multiple Lgistic Regressin The lg-rati between class k and l is lg Pr(Y = k X = x) Pr(Y = l X = x) = lg π k + lg g k(x) π l g l (x) Fr each class k, its discriminant functin is = [x T Σ µ k µt k Σ µ k + lg π k ] [x T Σ µ l µt l Σ µ l + lg π l ] f k (x) = x T Σ µ k µt k Σ µ k + lg π k, The quadratic term canceled due t the equal cvariance matrix. The decisin rule is Ŷ (x) = argmax k=,...,k f k (x). Ha Helen Zhang Lecture 8: Multiclass Classificatin (I)

27 Bayes Rule fr Multiclass Prblems Traditinal Methds fr Multiclass Prblems Linear Regressin Mdels Multivariate Linear Regressin Multiple Lgistic Regressin Ð Ñ ÒØ Ó ËØ Ø Ø Ð Ä ÖÒ Ò À Ø Ì Ö Ò ² Ö Ñ Ò ¾¼¼½ ÔØ Ö ÙÖ º Ì Ð Ø Ô Ò Ð ÓÛ Ø Ö Ù Ò ØÖ ¹ ÙØ ÓÒ Û Ø Ø Ñ ÓÚ Ö Ò Ò «Ö ÒØ Ñ Ò º ÁÒÐÙ Ö Ø ÓÒØÓÙÖ Ó ÓÒ Ø ÒØ Ò ØÝ ÒÐÓ Ò ± Ó Ø ÔÖÓ Ð ØÝ Ò º Ì Ý ÓÒ ÓÙÒ Ö ØÛ Ò Ô Ö Ó Ð Ö ÓÛÒ ÖÓ¹ Ò ØÖ Ø Ð Ò µ Ò Ø Ý ÓÒ ÓÙÒ Ö Ô Ö Ø Ò ÐÐ Ø Ö Ð Ö Ø Ø Ö ÓÐ Ð Ò Ù Ø Ó Ø ÓÖÑ Öµº ÇÒ Ø Ö Ø Û ÑÔÐ Ó ¼ Ö ÛÒ ÖÓÑ Ù Ò ØÖ ÙØ ÓÒ Ò Ø ØØ Ä ÓÒ ÓÙÒ Ö º Ha Helen Zhang Lecture 8: Multiclass Classificatin (I)

28 Bayes Rule fr Multiclass Prblems Traditinal Methds fr Multiclass Prblems Linear Regressin Mdels Interpretatin f LDA Multivariate Linear Regressin Multiple Lgistic Regressin lg Pr(Y = k X = x) = (x µ k) T Σ (x µ k ) + lg π k + cnst. The cnstant term des nt invlve x. If prir prbabilities are same, the LDA classifies x t the class with centrid clsest t x, using the squared Mahalanbis distance, d Σ (x, µ), based n the cmmn cvariance matrix. If d Σ (x µ k ) = d Σ (x µ l ), then the prir determines the classifier. Special Case: Σ k = I fr all k f k (x) = x µ k + lg π k. nly Euclidean distance is needed! Ha Helen Zhang Lecture 8: Multiclass Classificatin (I)

29 Bayes Rule fr Multiclass Prblems Traditinal Methds fr Multiclass Prblems Linear Regressin Mdels Mahalanbis Distance Multivariate Linear Regressin Multiple Lgistic Regressin Mahalanbis distance is a distance measure intrduced by P. C. Mahalanbis in 96. Def: The Mahalanbis distance f x = (x,..., x d ) T frm a set f pints with mean µ = (µ,..., µ d ) T and cvariance matrix Σ is defined as: [ / d Σ (x, µ) = (x µ) T Σ (x µ)]. It differs frm Euclidean distance in that it takes int accunt the crrelatins f the data set it is scale-invariant. It can als be defined as a dissimilarity measure between x and x f the same distributin with the cvariance matrix Σ: [ / d Σ (x, x ) = (x x ) T Σ (x x )]. Ha Helen Zhang Lecture 8: Multiclass Classificatin (I)

30 Bayes Rule fr Multiclass Prblems Traditinal Methds fr Multiclass Prblems Linear Regressin Mdels Special Cases Multivariate Linear Regressin Multiple Lgistic Regressin If Σ is the identity matrix, the Mahalanbis distance reduces t the Euclidean distance. If Σ is diagnal diag(σ,, σ d ), then the resulting distance measure is called the nrmalized Euclidean distance: d Σ (x, x ) = [ d i= ] / (x i x i ) σi, where σ i is the standard deviatin f the x i and x i. Ha Helen Zhang Lecture 8: Multiclass Classificatin (I)

31 Bayes Rule fr Multiclass Prblems Traditinal Methds fr Multiclass Prblems Linear Regressin Mdels Nrmalized Distance Multivariate Linear Regressin Multiple Lgistic Regressin Assume we like t decide whether x belngs t a class. Intuitively, the decisin relies n the distance f x t the class center µ: the clser t µ, the mre likely x belngs t the class. T decide whether a given distance is large r small, we need t knw if the pints in the class are spread ut ver a large range r a small range. Statistically, we measure the spread by the standard deviatin f distances f the sample pints t µ. If d Σ (x, µ) is less than ne standard deviatin, then it is highly prbable that x belngs t the class. The further away it is, the mre likely that x nt belnging t the class. This is the cncept f the nrmalized distance. Ha Helen Zhang Lecture 8: Multiclass Classificatin (I)

32 Bayes Rule fr Multiclass Prblems Traditinal Methds fr Multiclass Prblems Linear Regressin Mdels Mtivatin f Mahalanbis Distance Multivariate Linear Regressin Multiple Lgistic Regressin The drawback f the nrmalized distance is that the sample pints are assumed t be distributed abut µ in a spherical manner. Fr nn-spherical situatins, fr instance ellipsidal, the prbability f x belnging t the class depends nt nly n the distance frm µ, but als n the directin. In thse directins where the ellipsid has a shrt axis, x must be clser. In thse where the axis is lng, x can be further away frm the center. The ellipsid that best represents the class s prbability distributin can be estimated by the cvariance matrix f the samples. Ha Helen Zhang Lecture 8: Multiclass Classificatin (I)

33 Bayes Rule fr Multiclass Prblems Traditinal Methds fr Multiclass Prblems Linear Regressin Mdels Linear Decisin Bundary f LDA Linear discriminant functins Multivariate Linear Regressin Multiple Lgistic Regressin f k (x) = β T k x + β 0k, f j (x) = β T jx + β 0j. Decisin bundary functin between class k and j is (β k β j ) T x + (β 0k β 0j ) = β T x + β 0 = 0. Since β k = Σ µ k and β j = Σ µ j, the decisin bundary has the directinal vectr β = Σ (µ k µ j ) β in generally nt in the directin f µ k µ j The discriminant directin β minimizes the verlap fr Gaussian data Ha Helen Zhang Lecture 8: Multiclass Classificatin (I)

34 Bayes Rule fr Multiclass Prblems Traditinal Methds fr Multiclass Prblems Linear Regressin Mdels Multivariate Linear Regressin Multiple Lgistic Regressin Ð Ñ ÒØ Ó ËØ Ø Ø Ð Ä ÖÒ Ò À Ø Ì Ö Ò ² Ö Ñ Ò ¾¼¼½ ÔØ Ö ÙÖ º ÐØ ÓÙ Ø Ð Ò Ó Ò Ò Ø ÒØÖÓ Ò Ø Ö Ø ÓÒ Ó Ö Ø Ø ÒØÖÓ ÔÖ Ø ÔÖÓ Ø Ø ÓÚ ÖÐ Ô Ù Ó Ø ÓÚ Ö Ò Ð Ø Ô Ò Ðµº Ì Ö Ñ Ò ÒØ Ö Ø ÓÒ Ñ Ò Ñ Þ Ø ÓÚ ÖÐ Ô ÓÖ Ù Ò Ø Ö Ø Ô Ò Ðµº Ha Helen Zhang Lecture 8: Multiclass Classificatin (I)

35 Bayes Rule fr Multiclass Prblems Traditinal Methds fr Multiclass Prblems Linear Regressin Mdels Parameter Estimatin in LDA Multivariate Linear Regressin Multiple Lgistic Regressin In practice, we estimate the parameters frm the training data ˆπ k = n k /n, where n k is sample size in ˆµ k = Y i =k x i/n k The sample cvariance matrix fr the kth class: S k = n k (x i ˆµ k )(x i ˆµ k ) T. Y i =k (Unbiased) pled sample cvariance is a weighted average ˆΣ = = K k= n k K l= (n l ) S k K (x i ˆµ k )(x i ˆµ k ) T /(n K). k= Y i =k Ha Helen Zhang Lecture 8: Multiclass Classificatin (I)

36 Bayes Rule fr Multiclass Prblems Traditinal Methds fr Multiclass Prblems Linear Regressin Mdels Implemenatin f LDA Multivariate Linear Regressin Multiple Lgistic Regressin Cmpute the eigen-decmpsitin: ˆΣ = UDU T Sphere the data using ˆΣ, using the fllwing transfrmatin X = D / U T X. The cmmn cvariance estimate f X is the identity matrix. In the transfrmed space, classify the pint t the clsest transfrmed centrid, with the adjustment using π k s. Ha Helen Zhang Lecture 8: Multiclass Classificatin (I)

37 Bayes Rule fr Multiclass Prblems Traditinal Methds fr Multiclass Prblems Linear Regressin Mdels Multivariate Linear Regressin Multiple Lgistic Regressin Cnnectin between LDA and Linear Regressin fr K = Suppse we cde the targets in the tw classes as + and. Fit linear mdel y i = β 0 + β T x i, i =,, n Then the OLS ( ˆβ 0, β ) satisfy β ˆΣ ( µ µ ) ˆβ 0 is different frm the LDA unless n = n. The linear regressin gives a different decisin rule frm LDA, unless n = n. Ha Helen Zhang Lecture 8: Multiclass Classificatin (I)

38 Bayes Rule fr Multiclass Prblems Traditinal Methds fr Multiclass Prblems Linear Regressin Mdels Multiple Lgistic Mdels Multivariate Linear Regressin Multiple Lgistic Regressin Mdel the K psterir prbabilities by linear functins f x. We use lgit transfrmatins Pr(Y = X = x) lg = β 0 + β T x Pr(Y = K X = x) Pr(Y = X = x) lg = β 0 + β T x Pr(Y = K X = x) Pr(Y = K X = x) lg = β Pr(Y = K X = x) (K )0 + β T K x The chice f denminatr in the dd-ratis is arbitrary. The parameter vectr θ = {β 0, β,..., β (K )0, β K }. Ha Helen Zhang Lecture 8: Multiclass Classificatin (I)

39 Bayes Rule fr Multiclass Prblems Traditinal Methds fr Multiclass Prblems Linear Regressin Mdels Prbability Estimates Multivariate Linear Regressin Multiple Lgistic Regressin We use the lgit transfrmatin t assure that the ttal prbabilities sum t ne each estimated prbability lies in [0, ] p k (x) Pr(Y = k x) = exp (β k0 + β T k x) + K l= exp (β l0 + β T l x) fr k =,..., K. p K (x) Pr(Y = K x) = + K l= exp (β l0 + β T l x) Check K k= p k(x) = fr any x. Ha Helen Zhang Lecture 8: Multiclass Classificatin (I)

40 Bayes Rule fr Multiclass Prblems Traditinal Methds fr Multiclass Prblems Linear Regressin Mdels Multivariate Linear Regressin Multiple Lgistic Regressin Maximum Likelihd Estimate fr Lgistic Mdels The jint cnditinal likelihd f y i given x i is l(θ) = n lg p yi (θ; x i ) i= Using the Newtn-Raphsn methd: slve iteratively by minimizing re-weighted least squares. prvide cnsistent estimates if the mdels are specified crrectly. Ha Helen Zhang Lecture 8: Multiclass Classificatin (I)

COMP 551 Applied Machine Learning Lecture 5: Generative models for linear classification

COMP 551 Applied Machine Learning Lecture 5: Generative models for linear classification COMP 551 Applied Machine Learning Lecture 5: Generative mdels fr linear classificatin Instructr: Herke van Hf (herke.vanhf@mail.mcgill.ca) Slides mstly by: Jelle Pineau Class web page: www.cs.mcgill.ca/~hvanh2/cmp551

More information

Lecture 10, Principal Component Analysis

Lecture 10, Principal Component Analysis Principal Cmpnent Analysis Lecture 10, Principal Cmpnent Analysis Ha Helen Zhang Fall 2017 Ha Helen Zhang Lecture 10, Principal Cmpnent Analysis 1 / 16 Principal Cmpnent Analysis Lecture 10, Principal

More information

x 1 Outline IAML: Logistic Regression Decision Boundaries Example Data

x 1 Outline IAML: Logistic Regression Decision Boundaries Example Data Outline IAML: Lgistic Regressin Charles Suttn and Victr Lavrenk Schl f Infrmatics Semester Lgistic functin Lgistic regressin Learning lgistic regressin Optimizatin The pwer f nn-linear basis functins Least-squares

More information

Linear Classification

Linear Classification Linear Classificatin CS 54: Machine Learning Slides adapted frm Lee Cper, Jydeep Ghsh, and Sham Kakade Review: Linear Regressin CS 54 [Spring 07] - H Regressin Given an input vectr x T = (x, x,, xp), we

More information

Pattern Recognition 2014 Support Vector Machines

Pattern Recognition 2014 Support Vector Machines Pattern Recgnitin 2014 Supprt Vectr Machines Ad Feelders Universiteit Utrecht Ad Feelders ( Universiteit Utrecht ) Pattern Recgnitin 1 / 55 Overview 1 Separable Case 2 Kernel Functins 3 Allwing Errrs (Sft

More information

Resampling Methods. Cross-validation, Bootstrapping. Marek Petrik 2/21/2017

Resampling Methods. Cross-validation, Bootstrapping. Marek Petrik 2/21/2017 Resampling Methds Crss-validatin, Btstrapping Marek Petrik 2/21/2017 Sme f the figures in this presentatin are taken frm An Intrductin t Statistical Learning, with applicatins in R (Springer, 2013) with

More information

COMP 551 Applied Machine Learning Lecture 4: Linear classification

COMP 551 Applied Machine Learning Lecture 4: Linear classification COMP 551 Applied Machine Learning Lecture 4: Linear classificatin Instructr: Jelle Pineau (jpineau@cs.mcgill.ca) Class web page: www.cs.mcgill.ca/~jpineau/cmp551 Unless therwise nted, all material psted

More information

Lecture 2: Supervised vs. unsupervised learning, bias-variance tradeoff

Lecture 2: Supervised vs. unsupervised learning, bias-variance tradeoff Lecture 2: Supervised vs. unsupervised learning, bias-variance tradeff Reading: Chapter 2 STATS 202: Data mining and analysis September 27, 2017 1 / 20 Supervised vs. unsupervised learning In unsupervised

More information

Lecture 2: Supervised vs. unsupervised learning, bias-variance tradeoff

Lecture 2: Supervised vs. unsupervised learning, bias-variance tradeoff Lecture 2: Supervised vs. unsupervised learning, bias-variance tradeff Reading: Chapter 2 STATS 202: Data mining and analysis September 27, 2017 1 / 20 Supervised vs. unsupervised learning In unsupervised

More information

Stats Classification Ji Zhu, Michigan Statistics 1. Classification. Ji Zhu 445C West Hall

Stats Classification Ji Zhu, Michigan Statistics 1. Classification. Ji Zhu 445C West Hall Stats 415 - Classificatin Ji Zhu, Michigan Statistics 1 Classificatin Ji Zhu 445C West Hall 734-936-2577 jizhu@umich.edu Stats 415 - Classificatin Ji Zhu, Michigan Statistics 2 Examples f Classificatin

More information

Part 3 Introduction to statistical classification techniques

Part 3 Introduction to statistical classification techniques Part 3 Intrductin t statistical classificatin techniques Machine Learning, Part 3, March 07 Fabi Rli Preamble ØIn Part we have seen that if we knw: Psterir prbabilities P(ω i / ) Or the equivalent terms

More information

Distributions, spatial statistics and a Bayesian perspective

Distributions, spatial statistics and a Bayesian perspective Distributins, spatial statistics and a Bayesian perspective Dug Nychka Natinal Center fr Atmspheric Research Distributins and densities Cnditinal distributins and Bayes Thm Bivariate nrmal Spatial statistics

More information

Midwest Big Data Summer School: Machine Learning I: Introduction. Kris De Brabanter

Midwest Big Data Summer School: Machine Learning I: Introduction. Kris De Brabanter Midwest Big Data Summer Schl: Machine Learning I: Intrductin Kris De Brabanter kbrabant@iastate.edu Iwa State University Department f Statistics Department f Cmputer Science June 24, 2016 1/24 Outline

More information

Resampling Methods. Chapter 5. Chapter 5 1 / 52

Resampling Methods. Chapter 5. Chapter 5 1 / 52 Resampling Methds Chapter 5 Chapter 5 1 / 52 1 51 Validatin set apprach 2 52 Crss validatin 3 53 Btstrap Chapter 5 2 / 52 Abut Resampling An imprtant statistical tl Pretending the data as ppulatin and

More information

Support-Vector Machines

Support-Vector Machines Supprt-Vectr Machines Intrductin Supprt vectr machine is a linear machine with sme very nice prperties. Haykin chapter 6. See Alpaydin chapter 13 fr similar cntent. Nte: Part f this lecture drew material

More information

COMP 551 Applied Machine Learning Lecture 11: Support Vector Machines

COMP 551 Applied Machine Learning Lecture 11: Support Vector Machines COMP 551 Applied Machine Learning Lecture 11: Supprt Vectr Machines Instructr: (jpineau@cs.mcgill.ca) Class web page: www.cs.mcgill.ca/~jpineau/cmp551 Unless therwise nted, all material psted fr this curse

More information

CHAPTER 24: INFERENCE IN REGRESSION. Chapter 24: Make inferences about the population from which the sample data came.

CHAPTER 24: INFERENCE IN REGRESSION. Chapter 24: Make inferences about the population from which the sample data came. MATH 1342 Ch. 24 April 25 and 27, 2013 Page 1 f 5 CHAPTER 24: INFERENCE IN REGRESSION Chapters 4 and 5: Relatinships between tw quantitative variables. Be able t Make a graph (scatterplt) Summarize the

More information

Support Vector Machines and Flexible Discriminants

Support Vector Machines and Flexible Discriminants 12 Supprt Vectr Machines and Flexible Discriminants This is page 417 Printer: Opaque this 12.1 Intrductin In this chapter we describe generalizatins f linear decisin bundaries fr classificatin. Optimal

More information

Module 3: Gaussian Process Parameter Estimation, Prediction Uncertainty, and Diagnostics

Module 3: Gaussian Process Parameter Estimation, Prediction Uncertainty, and Diagnostics Mdule 3: Gaussian Prcess Parameter Estimatin, Predictin Uncertainty, and Diagnstics Jerme Sacks and William J Welch Natinal Institute f Statistical Sciences and University f British Clumbia Adapted frm

More information

k-nearest Neighbor How to choose k Average of k points more reliable when: Large k: noise in attributes +o o noise in class labels

k-nearest Neighbor How to choose k Average of k points more reliable when: Large k: noise in attributes +o o noise in class labels Mtivating Example Memry-Based Learning Instance-Based Learning K-earest eighbr Inductive Assumptin Similar inputs map t similar utputs If nt true => learning is impssible If true => learning reduces t

More information

A Matrix Representation of Panel Data

A Matrix Representation of Panel Data web Extensin 6 Appendix 6.A A Matrix Representatin f Panel Data Panel data mdels cme in tw brad varieties, distinct intercept DGPs and errr cmpnent DGPs. his appendix presents matrix algebra representatins

More information

Lecture 16: Modern Classification (I) - Separating Hyperplanes

Lecture 16: Modern Classification (I) - Separating Hyperplanes Lecture 16: Modern Classification (I) - Separating Hyperplanes Outline 1 2 Separating Hyperplane Binary SVM for Separable Case Bayes Rule for Binary Problems Consider the simplest case: two classes are

More information

3.4 Shrinkage Methods Prostate Cancer Data Example (Continued) Ridge Regression

3.4 Shrinkage Methods Prostate Cancer Data Example (Continued) Ridge Regression 3.3.4 Prstate Cancer Data Example (Cntinued) 3.4 Shrinkage Methds 61 Table 3.3 shws the cefficients frm a number f different selectin and shrinkage methds. They are best-subset selectin using an all-subsets

More information

What is Statistical Learning?

What is Statistical Learning? What is Statistical Learning? Sales 5 10 15 20 25 Sales 5 10 15 20 25 Sales 5 10 15 20 25 0 50 100 200 300 TV 0 10 20 30 40 50 Radi 0 20 40 60 80 100 Newspaper Shwn are Sales vs TV, Radi and Newspaper,

More information

Bootstrap Method > # Purpose: understand how bootstrap method works > obs=c(11.96, 5.03, 67.40, 16.07, 31.50, 7.73, 11.10, 22.38) > n=length(obs) >

Bootstrap Method > # Purpose: understand how bootstrap method works > obs=c(11.96, 5.03, 67.40, 16.07, 31.50, 7.73, 11.10, 22.38) > n=length(obs) > Btstrap Methd > # Purpse: understand hw btstrap methd wrks > bs=c(11.96, 5.03, 67.40, 16.07, 31.50, 7.73, 11.10, 22.38) > n=length(bs) > mean(bs) [1] 21.64625 > # estimate f lambda > lambda = 1/mean(bs);

More information

Simple Linear Regression (single variable)

Simple Linear Regression (single variable) Simple Linear Regressin (single variable) Intrductin t Machine Learning Marek Petrik January 31, 2017 Sme f the figures in this presentatin are taken frm An Intrductin t Statistical Learning, with applicatins

More information

IAML: Support Vector Machines

IAML: Support Vector Machines 1 / 22 IAML: Supprt Vectr Machines Charles Suttn and Victr Lavrenk Schl f Infrmatics Semester 1 2 / 22 Outline Separating hyperplane with maimum margin Nn-separable training data Epanding the input int

More information

Support Vector Machines and Flexible Discriminants

Support Vector Machines and Flexible Discriminants Supprt Vectr Machines and Flexible Discriminants This is page Printer: Opaque this. Intrductin In this chapter we describe generalizatins f linear decisin bundaries fr classificatin. Optimal separating

More information

Comparing Several Means: ANOVA. Group Means and Grand Mean

Comparing Several Means: ANOVA. Group Means and Grand Mean STAT 511 ANOVA and Regressin 1 Cmparing Several Means: ANOVA Slide 1 Blue Lake snap beans were grwn in 12 pen-tp chambers which are subject t 4 treatments 3 each with O 3 and SO 2 present/absent. The ttal

More information

Contents. This is page i Printer: Opaque this

Contents. This is page i Printer: Opaque this Cntents This is page i Printer: Opaque this Supprt Vectr Machines and Flexible Discriminants. Intrductin............. The Supprt Vectr Classifier.... Cmputing the Supprt Vectr Classifier........ Mixture

More information

SUPPLEMENTARY MATERIAL GaGa: a simple and flexible hierarchical model for microarray data analysis

SUPPLEMENTARY MATERIAL GaGa: a simple and flexible hierarchical model for microarray data analysis SUPPLEMENTARY MATERIAL GaGa: a simple and flexible hierarchical mdel fr micrarray data analysis David Rssell Department f Bistatistics M.D. Andersn Cancer Center, Hustn, TX 77030, USA rsselldavid@gmail.cm

More information

In SMV I. IAML: Support Vector Machines II. This Time. The SVM optimization problem. We saw:

In SMV I. IAML: Support Vector Machines II. This Time. The SVM optimization problem. We saw: In SMV I IAML: Supprt Vectr Machines II Nigel Gddard Schl f Infrmatics Semester 1 We sa: Ma margin trick Gemetry f the margin and h t cmpute it Finding the ma margin hyperplane using a cnstrained ptimizatin

More information

Internal vs. external validity. External validity. This section is based on Stock and Watson s Chapter 9.

Internal vs. external validity. External validity. This section is based on Stock and Watson s Chapter 9. Sectin 7 Mdel Assessment This sectin is based n Stck and Watsn s Chapter 9. Internal vs. external validity Internal validity refers t whether the analysis is valid fr the ppulatin and sample being studied.

More information

Tree Structured Classifier

Tree Structured Classifier Tree Structured Classifier Reference: Classificatin and Regressin Trees by L. Breiman, J. H. Friedman, R. A. Olshen, and C. J. Stne, Chapman & Hall, 98. A Medical Eample (CART): Predict high risk patients

More information

Lecture 3: Principal Components Analysis (PCA)

Lecture 3: Principal Components Analysis (PCA) Lecture 3: Principal Cmpnents Analysis (PCA) Reading: Sectins 6.3.1, 10.1, 10.2, 10.4 STATS 202: Data mining and analysis Jnathan Taylr, 9/28 Slide credits: Sergi Bacallad 1 / 24 The bias variance decmpsitin

More information

Computational modeling techniques

Computational modeling techniques Cmputatinal mdeling techniques Lecture 4: Mdel checing fr ODE mdels In Petre Department f IT, Åb Aademi http://www.users.ab.fi/ipetre/cmpmd/ Cntent Stichimetric matrix Calculating the mass cnservatin relatins

More information

COMP 551 Applied Machine Learning Lecture 9: Support Vector Machines (cont d)

COMP 551 Applied Machine Learning Lecture 9: Support Vector Machines (cont d) COMP 551 Applied Machine Learning Lecture 9: Supprt Vectr Machines (cnt d) Instructr: Herke van Hf (herke.vanhf@mail.mcgill.ca) Slides mstly by: Class web page: www.cs.mcgill.ca/~hvanh2/cmp551 Unless therwise

More information

Linear programming III

Linear programming III Linear prgramming III Review 1/33 What have cvered in previus tw classes LP prblem setup: linear bjective functin, linear cnstraints. exist extreme pint ptimal slutin. Simplex methd: g thrugh extreme pint

More information

Multiple Source Multiple. using Network Coding

Multiple Source Multiple. using Network Coding Multiple Surce Multiple Destinatin Tplgy Inference using Netwrk Cding Pegah Sattari EECS, UC Irvine Jint wrk with Athina Markpulu, at UCI, Christina Fraguli, at EPFL, Lausanne Outline Netwrk Tmgraphy Gal,

More information

initially lcated away frm the data set never win the cmpetitin, resulting in a nnptimal nal cdebk, [2] [3] [4] and [5]. Khnen's Self Organizing Featur

initially lcated away frm the data set never win the cmpetitin, resulting in a nnptimal nal cdebk, [2] [3] [4] and [5]. Khnen's Self Organizing Featur Cdewrd Distributin fr Frequency Sensitive Cmpetitive Learning with One Dimensinal Input Data Aristides S. Galanpuls and Stanley C. Ahalt Department f Electrical Engineering The Ohi State University Abstract

More information

Statistical classifiers: Bayesian decision theory and density estimation

Statistical classifiers: Bayesian decision theory and density estimation 3 rd NOSE Shrt Curse Alpbach, st 6 th Mar 004 Statistical classifiers: Bayesian decisin thery and density estimatin Ricard Gutierrez- Department f Cmputer Science rgutier@cs.tamu.edu http://research.cs.tamu.edu/prism

More information

The blessing of dimensionality for kernel methods

The blessing of dimensionality for kernel methods fr kernel methds Building classifiers in high dimensinal space Pierre Dupnt Pierre.Dupnt@ucluvain.be Classifiers define decisin surfaces in sme feature space where the data is either initially represented

More information

Overview of Supervised Learning

Overview of Supervised Learning 2 Overview f Supervised Learning 2.1 Intrductin The first three examples described in Chapter 1 have several cmpnents in cmmn. Fr each there is a set f variables that might be dented as inputs, which are

More information

CS 109 Lecture 23 May 18th, 2016

CS 109 Lecture 23 May 18th, 2016 CS 109 Lecture 23 May 18th, 2016 New Datasets Heart Ancestry Netflix Our Path Parameter Estimatin Machine Learning: Frmally Many different frms f Machine Learning We fcus n the prblem f predictin Want

More information

LHS Mathematics Department Honors Pre-Calculus Final Exam 2002 Answers

LHS Mathematics Department Honors Pre-Calculus Final Exam 2002 Answers LHS Mathematics Department Hnrs Pre-alculus Final Eam nswers Part Shrt Prblems The table at the right gives the ppulatin f Massachusetts ver the past several decades Using an epnential mdel, predict the

More information

The Kullback-Leibler Kernel as a Framework for Discriminant and Localized Representations for Visual Recognition

The Kullback-Leibler Kernel as a Framework for Discriminant and Localized Representations for Visual Recognition The Kullback-Leibler Kernel as a Framewrk fr Discriminant and Lcalized Representatins fr Visual Recgnitin Nun Vascncels Purdy H Pedr Mren ECE Department University f Califrnia, San Dieg HP Labs Cambridge

More information

Checking the resolved resonance region in EXFOR database

Checking the resolved resonance region in EXFOR database Checking the reslved resnance regin in EXFOR database Gttfried Bertn Sciété de Calcul Mathématique (SCM) Oscar Cabells OECD/NEA Data Bank JEFF Meetings - Sessin JEFF Experiments Nvember 0-4, 017 Bulgne-Billancurt,

More information

7 TH GRADE MATH STANDARDS

7 TH GRADE MATH STANDARDS ALGEBRA STANDARDS Gal 1: Students will use the language f algebra t explre, describe, represent, and analyze number expressins and relatins 7 TH GRADE MATH STANDARDS 7.M.1.1: (Cmprehensin) Select, use,

More information

Chapter 3: Cluster Analysis

Chapter 3: Cluster Analysis Chapter 3: Cluster Analysis } 3.1 Basic Cncepts f Clustering 3.1.1 Cluster Analysis 3.1. Clustering Categries } 3. Partitining Methds 3..1 The principle 3.. K-Means Methd 3..3 K-Medids Methd 3..4 CLARA

More information

Differentiation Applications 1: Related Rates

Differentiation Applications 1: Related Rates Differentiatin Applicatins 1: Related Rates 151 Differentiatin Applicatins 1: Related Rates Mdel 1: Sliding Ladder 10 ladder y 10 ladder 10 ladder A 10 ft ladder is leaning against a wall when the bttm

More information

, which yields. where z1. and z2

, which yields. where z1. and z2 The Gaussian r Nrmal PDF, Page 1 The Gaussian r Nrmal Prbability Density Functin Authr: Jhn M Cimbala, Penn State University Latest revisin: 11 September 13 The Gaussian r Nrmal Prbability Density Functin

More information

Slide04 (supplemental) Haykin Chapter 4 (both 2nd and 3rd ed): Multi-Layer Perceptrons

Slide04 (supplemental) Haykin Chapter 4 (both 2nd and 3rd ed): Multi-Layer Perceptrons Slide04 supplemental) Haykin Chapter 4 bth 2nd and 3rd ed): Multi-Layer Perceptrns CPSC 636-600 Instructr: Ynsuck Che Heuristic fr Making Backprp Perfrm Better 1. Sequential vs. batch update: fr large

More information

SURVIVAL ANALYSIS WITH SUPPORT VECTOR MACHINES

SURVIVAL ANALYSIS WITH SUPPORT VECTOR MACHINES 1 SURVIVAL ANALYSIS WITH SUPPORT VECTOR MACHINES Wlfgang HÄRDLE Ruslan MORO Center fr Applied Statistics and Ecnmics (CASE), Humbldt-Universität zu Berlin Mtivatin 2 Applicatins in Medicine estimatin f

More information

Reinforcement Learning" CMPSCI 383 Nov 29, 2011!

Reinforcement Learning CMPSCI 383 Nov 29, 2011! Reinfrcement Learning" CMPSCI 383 Nv 29, 2011! 1 Tdayʼs lecture" Review f Chapter 17: Making Cmple Decisins! Sequential decisin prblems! The mtivatin and advantages f reinfrcement learning.! Passive learning!

More information

Principal Components

Principal Components Principal Cmpnents Suppse we have N measurements n each f p variables X j, j = 1,..., p. There are several equivalent appraches t principal cmpnents: Given X = (X 1,... X p ), prduce a derived (and small)

More information

CMSC858P Supervised Learning Methods

CMSC858P Supervised Learning Methods CMSC858P Supervised Learning Methods Hector Corrada Bravo March, 2010 Introduction Today we discuss the classification setting in detail. Our setting is that we observe for each subject i a set of p predictors

More information

MATHEMATICS SYLLABUS SECONDARY 5th YEAR

MATHEMATICS SYLLABUS SECONDARY 5th YEAR Eurpean Schls Office f the Secretary-General Pedaggical Develpment Unit Ref. : 011-01-D-8-en- Orig. : EN MATHEMATICS SYLLABUS SECONDARY 5th YEAR 6 perid/week curse APPROVED BY THE JOINT TEACHING COMMITTEE

More information

AP Statistics Practice Test Unit Three Exploring Relationships Between Variables. Name Period Date

AP Statistics Practice Test Unit Three Exploring Relationships Between Variables. Name Period Date AP Statistics Practice Test Unit Three Explring Relatinships Between Variables Name Perid Date True r False: 1. Crrelatin and regressin require explanatry and respnse variables. 1. 2. Every least squares

More information

Elements of Machine Intelligence - I

Elements of Machine Intelligence - I ECE-175A Elements f Machine Intelligence - I Ken Kreutz-Delgad Nun Vascncels ECE Department, UCSD Winter 2011 The curse The curse will cver basic, but imprtant, aspects f machine learning and pattern recgnitin

More information

AP Statistics Notes Unit Two: The Normal Distributions

AP Statistics Notes Unit Two: The Normal Distributions AP Statistics Ntes Unit Tw: The Nrmal Distributins Syllabus Objectives: 1.5 The student will summarize distributins f data measuring the psitin using quartiles, percentiles, and standardized scres (z-scres).

More information

STATS216v Introduction to Statistical Learning Stanford University, Summer Practice Final (Solutions) Duration: 3 hours

STATS216v Introduction to Statistical Learning Stanford University, Summer Practice Final (Solutions) Duration: 3 hours STATS216v Intrductin t Statistical Learning Stanfrd University, Summer 2016 Practice Final (Slutins) Duratin: 3 hurs Instructins: (This is a practice final and will nt be graded.) Remember the university

More information

Lead/Lag Compensator Frequency Domain Properties and Design Methods

Lead/Lag Compensator Frequency Domain Properties and Design Methods Lectures 6 and 7 Lead/Lag Cmpensatr Frequency Dmain Prperties and Design Methds Definitin Cnsider the cmpensatr (ie cntrller Fr, it is called a lag cmpensatr s K Fr s, it is called a lead cmpensatr Ntatin

More information

T Algorithmic methods for data mining. Slide set 6: dimensionality reduction

T Algorithmic methods for data mining. Slide set 6: dimensionality reduction T-61.5060 Algrithmic methds fr data mining Slide set 6: dimensinality reductin reading assignment LRU bk: 11.1 11.3 PCA tutrial in mycurses (ptinal) ptinal: An Elementary Prf f a Therem f Jhnsn and Lindenstrauss,

More information

NOTE ON THE ANALYSIS OF A RANDOMIZED BLOCK DESIGN. Junjiro Ogawa University of North Carolina

NOTE ON THE ANALYSIS OF A RANDOMIZED BLOCK DESIGN. Junjiro Ogawa University of North Carolina NOTE ON THE ANALYSIS OF A RANDOMIZED BLOCK DESIGN by Junjir Ogawa University f Nrth Carlina This research was supprted by the Office f Naval Research under Cntract N. Nnr-855(06) fr research in prbability

More information

CN700 Additive Models and Trees Chapter 9: Hastie et al. (2001)

CN700 Additive Models and Trees Chapter 9: Hastie et al. (2001) CN700 Additive Mdels and Trees Chapter 9: Hastie et al. (2001) Madhusudana Shashanka Department f Cgnitive and Neural Systems Bstn University CN700 - Additive Mdels and Trees March 02, 2004 p.1/34 Overview

More information

CHAPTER 3 INEQUALITIES. Copyright -The Institute of Chartered Accountants of India

CHAPTER 3 INEQUALITIES. Copyright -The Institute of Chartered Accountants of India CHAPTER 3 INEQUALITIES Cpyright -The Institute f Chartered Accuntants f India INEQUALITIES LEARNING OBJECTIVES One f the widely used decisin making prblems, nwadays, is t decide n the ptimal mix f scarce

More information

The general linear model and Statistical Parametric Mapping I: Introduction to the GLM

The general linear model and Statistical Parametric Mapping I: Introduction to the GLM The general linear mdel and Statistical Parametric Mapping I: Intrductin t the GLM Alexa Mrcm and Stefan Kiebel, Rik Hensn, Andrew Hlmes & J-B J Pline Overview Intrductin Essential cncepts Mdelling Design

More information

Fall 2013 Physics 172 Recitation 3 Momentum and Springs

Fall 2013 Physics 172 Recitation 3 Momentum and Springs Fall 03 Physics 7 Recitatin 3 Mmentum and Springs Purpse: The purpse f this recitatin is t give yu experience wrking with mmentum and the mmentum update frmula. Readings: Chapter.3-.5 Learning Objectives:.3.

More information

A New Evaluation Measure. J. Joiner and L. Werner. The problems of evaluation and the needed criteria of evaluation

A New Evaluation Measure. J. Joiner and L. Werner. The problems of evaluation and the needed criteria of evaluation III-l III. A New Evaluatin Measure J. Jiner and L. Werner Abstract The prblems f evaluatin and the needed criteria f evaluatin measures in the SMART system f infrmatin retrieval are reviewed and discussed.

More information

Building to Transformations on Coordinate Axis Grade 5: Geometry Graph points on the coordinate plane to solve real-world and mathematical problems.

Building to Transformations on Coordinate Axis Grade 5: Geometry Graph points on the coordinate plane to solve real-world and mathematical problems. Building t Transfrmatins n Crdinate Axis Grade 5: Gemetry Graph pints n the crdinate plane t slve real-wrld and mathematical prblems. 5.G.1. Use a pair f perpendicular number lines, called axes, t define

More information

PSU GISPOPSCI June 2011 Ordinary Least Squares & Spatial Linear Regression in GeoDa

PSU GISPOPSCI June 2011 Ordinary Least Squares & Spatial Linear Regression in GeoDa There are tw parts t this lab. The first is intended t demnstrate hw t request and interpret the spatial diagnstics f a standard OLS regressin mdel using GeDa. The diagnstics prvide infrmatin abut the

More information

Math Foundations 20 Work Plan

Math Foundations 20 Work Plan Math Fundatins 20 Wrk Plan Units / Tpics 20.8 Demnstrate understanding f systems f linear inequalities in tw variables. Time Frame December 1-3 weeks 6-10 Majr Learning Indicatrs Identify situatins relevant

More information

Admin. MDP Search Trees. Optimal Quantities. Reinforcement Learning

Admin. MDP Search Trees. Optimal Quantities. Reinforcement Learning Admin Reinfrcement Learning Cntent adapted frm Berkeley CS188 MDP Search Trees Each MDP state prjects an expectimax-like search tree Optimal Quantities The value (utility) f a state s: V*(s) = expected

More information

More Tutorial at

More Tutorial at Answer each questin in the space prvided; use back f page if extra space is needed. Answer questins s the grader can READILY understand yur wrk; nly wrk n the exam sheet will be cnsidered. Write answers,

More information

Smoothing, penalized least squares and splines

Smoothing, penalized least squares and splines Smthing, penalized least squares and splines Duglas Nychka, www.image.ucar.edu/~nychka Lcally weighted averages Penalized least squares smthers Prperties f smthers Splines and Reprducing Kernels The interplatin

More information

4th Indian Institute of Astrophysics - PennState Astrostatistics School July, 2013 Vainu Bappu Observatory, Kavalur. Correlation and Regression

4th Indian Institute of Astrophysics - PennState Astrostatistics School July, 2013 Vainu Bappu Observatory, Kavalur. Correlation and Regression 4th Indian Institute f Astrphysics - PennState Astrstatistics Schl July, 2013 Vainu Bappu Observatry, Kavalur Crrelatin and Regressin Rahul Ry Indian Statistical Institute, Delhi. Crrelatin Cnsider a tw

More information

Lecture 5: Equilibrium and Oscillations

Lecture 5: Equilibrium and Oscillations Lecture 5: Equilibrium and Oscillatins Energy and Mtin Last time, we fund that fr a system with energy cnserved, v = ± E U m ( ) ( ) One result we see immediately is that there is n slutin fr velcity if

More information

Lecture 17: Free Energy of Multi-phase Solutions at Equilibrium

Lecture 17: Free Energy of Multi-phase Solutions at Equilibrium Lecture 17: 11.07.05 Free Energy f Multi-phase Slutins at Equilibrium Tday: LAST TIME...2 FREE ENERGY DIAGRAMS OF MULTI-PHASE SOLUTIONS 1...3 The cmmn tangent cnstructin and the lever rule...3 Practical

More information

ECEN 4872/5827 Lecture Notes

ECEN 4872/5827 Lecture Notes ECEN 4872/5827 Lecture Ntes Lecture #5 Objectives fr lecture #5: 1. Analysis f precisin current reference 2. Appraches fr evaluating tlerances 3. Temperature Cefficients evaluatin technique 4. Fundamentals

More information

Equilibrium of Stress

Equilibrium of Stress Equilibrium f Stress Cnsider tw perpendicular planes passing thrugh a pint p. The stress cmpnents acting n these planes are as shwn in ig. 3.4.1a. These stresses are usuall shwn tgether acting n a small

More information

Logistic Regression. and Maximum Likelihood. Marek Petrik. Feb

Logistic Regression. and Maximum Likelihood. Marek Petrik. Feb Lgistic Regressin and Maximum Likelihd Marek Petrik Feb 09 2017 S Far in ML Regressin vs Classificatin Linear regressin Bias-variance decmpsitin Practical methds fr linear regressin Simple Linear Regressin

More information

Biplots in Practice MICHAEL GREENACRE. Professor of Statistics at the Pompeu Fabra University. Chapter 13 Offprint

Biplots in Practice MICHAEL GREENACRE. Professor of Statistics at the Pompeu Fabra University. Chapter 13 Offprint Biplts in Practice MICHAEL GREENACRE Prfessr f Statistics at the Pmpeu Fabra University Chapter 13 Offprint CASE STUDY BIOMEDICINE Cmparing Cancer Types Accrding t Gene Epressin Arrays First published:

More information

Inference in the Multiple-Regression

Inference in the Multiple-Regression Sectin 5 Mdel Inference in the Multiple-Regressin Kinds f hypthesis tests in a multiple regressin There are several distinct kinds f hypthesis tests we can run in a multiple regressin. Suppse that amng

More information

Maximum A Posteriori (MAP) CS 109 Lecture 22 May 16th, 2016

Maximum A Posteriori (MAP) CS 109 Lecture 22 May 16th, 2016 Maximum A Psteriri (MAP) CS 109 Lecture 22 May 16th, 2016 Previusly in CS109 Game f Estimatrs Maximum Likelihd Nn spiler: this didn t happen Side Plt argmax argmax f lg Mther f ptimizatins? Reviving an

More information

MODULE FOUR. This module addresses functions. SC Academic Elementary Algebra Standards:

MODULE FOUR. This module addresses functions. SC Academic Elementary Algebra Standards: MODULE FOUR This mdule addresses functins SC Academic Standards: EA-3.1 Classify a relatinship as being either a functin r nt a functin when given data as a table, set f rdered pairs, r graph. EA-3.2 Use

More information

[COLLEGE ALGEBRA EXAM I REVIEW TOPICS] ( u s e t h i s t o m a k e s u r e y o u a r e r e a d y )

[COLLEGE ALGEBRA EXAM I REVIEW TOPICS] ( u s e t h i s t o m a k e s u r e y o u a r e r e a d y ) (Abut the final) [COLLEGE ALGEBRA EXAM I REVIEW TOPICS] ( u s e t h i s t m a k e s u r e y u a r e r e a d y ) The department writes the final exam s I dn't really knw what's n it and I can't very well

More information

CS 477/677 Analysis of Algorithms Fall 2007 Dr. George Bebis Course Project Due Date: 11/29/2007

CS 477/677 Analysis of Algorithms Fall 2007 Dr. George Bebis Course Project Due Date: 11/29/2007 CS 477/677 Analysis f Algrithms Fall 2007 Dr. Gerge Bebis Curse Prject Due Date: 11/29/2007 Part1: Cmparisn f Srting Algrithms (70% f the prject grade) The bjective f the first part f the assignment is

More information

8. Issues with Bases

8. Issues with Bases 8. Issues with Bases Nnparametric regressin ften tries t fit a mdel f the frm f(x) = M j=1 β j h j (x) + ǫ where the h j functins may pick ut specific cmpnents f x (multiple linear regressin), r be pwers

More information

Lecture 18: Multiclass Support Vector Machines

Lecture 18: Multiclass Support Vector Machines Fall, 2017 Outlines Overview of Multiclass Learning Traditional Methods for Multiclass Problems One-vs-rest approaches Pairwise approaches Recent development for Multiclass Problems Simultaneous Classification

More information

Computational modeling techniques

Computational modeling techniques Cmputatinal mdeling techniques Lecture 11: Mdeling with systems f ODEs In Petre Department f IT, Ab Akademi http://www.users.ab.fi/ipetre/cmpmd/ Mdeling with differential equatins Mdeling strategy Fcus

More information

Versatility of Singular Value Decomposition (SVD) January 7, 2015

Versatility of Singular Value Decomposition (SVD) January 7, 2015 Versatility f Singular Value Decmpsitin (SVD) January 7, 2015 Assumptin : Data = Real Data + Nise Each Data Pint is a clumn f the n d Data Matrix A. Assumptin : Data = Real Data + Nise Each Data Pint is

More information

UNIV1"'RSITY OF NORTH CAROLINA Department of Statistics Chapel Hill, N. C. CUMULATIVE SUM CONTROL CHARTS FOR THE FOLDED NORMAL DISTRIBUTION

UNIV1'RSITY OF NORTH CAROLINA Department of Statistics Chapel Hill, N. C. CUMULATIVE SUM CONTROL CHARTS FOR THE FOLDED NORMAL DISTRIBUTION UNIV1"'RSITY OF NORTH CAROLINA Department f Statistics Chapel Hill, N. C. CUMULATIVE SUM CONTROL CHARTS FOR THE FOLDED NORMAL DISTRIBUTION by N. L. Jlmsn December 1962 Grant N. AFOSR -62..148 Methds f

More information

24 Multiple Eigenvectors; Latent Factor Analysis; Nearest Neighbors

24 Multiple Eigenvectors; Latent Factor Analysis; Nearest Neighbors Multiple Eigenvectrs; Latent Factr Analysis; Nearest Neighbrs 47 24 Multiple Eigenvectrs; Latent Factr Analysis; Nearest Neighbrs Clustering w/multiple Eigenvectrs [When we use the Fiedler vectr fr spectral

More information

MATCHING TECHNIQUES. Technical Track Session VI. Emanuela Galasso. The World Bank

MATCHING TECHNIQUES. Technical Track Session VI. Emanuela Galasso. The World Bank MATCHING TECHNIQUES Technical Track Sessin VI Emanuela Galass The Wrld Bank These slides were develped by Christel Vermeersch and mdified by Emanuela Galass fr the purpse f this wrkshp When can we use

More information

MODULE 1. e x + c. [You can t separate a demominator, but you can divide a single denominator into each numerator term] a + b a(a + b)+1 = a + b

MODULE 1. e x + c. [You can t separate a demominator, but you can divide a single denominator into each numerator term] a + b a(a + b)+1 = a + b . REVIEW OF SOME BASIC ALGEBRA MODULE () Slving Equatins Yu shuld be able t slve fr x: a + b = c a d + e x + c and get x = e(ba +) b(c a) d(ba +) c Cmmn mistakes and strategies:. a b + c a b + a c, but

More information

Comparison of hybrid ensemble-4dvar with EnKF and 4DVar for regional-scale data assimilation

Comparison of hybrid ensemble-4dvar with EnKF and 4DVar for regional-scale data assimilation Cmparisn f hybrid ensemble-4dvar with EnKF and 4DVar fr reginal-scale data assimilatin Jn Pterjy and Fuqing Zhang Department f Meterlgy The Pennsylvania State University Wednesday 18 th December, 2013

More information

Eric Klein and Ning Sa

Eric Klein and Ning Sa Week 12. Statistical Appraches t Netwrks: p1 and p* Wasserman and Faust Chapter 15: Statistical Analysis f Single Relatinal Netwrks There are fur tasks in psitinal analysis: 1) Define Equivalence 2) Measure

More information

IN a recent article, Geary [1972] discussed the merit of taking first differences

IN a recent article, Geary [1972] discussed the merit of taking first differences The Efficiency f Taking First Differences in Regressin Analysis: A Nte J. A. TILLMAN IN a recent article, Geary [1972] discussed the merit f taking first differences t deal with the prblems that trends

More information

A Correlation of. to the. South Carolina Academic Standards for Mathematics Precalculus

A Correlation of. to the. South Carolina Academic Standards for Mathematics Precalculus A Crrelatin f Suth Carlina Academic Standards fr Mathematics Precalculus INTRODUCTION This dcument demnstrates hw Precalculus (Blitzer), 4 th Editin 010, meets the indicatrs f the. Crrelatin page references

More information

CAUSAL INFERENCE. Technical Track Session I. Phillippe Leite. The World Bank

CAUSAL INFERENCE. Technical Track Session I. Phillippe Leite. The World Bank CAUSAL INFERENCE Technical Track Sessin I Phillippe Leite The Wrld Bank These slides were develped by Christel Vermeersch and mdified by Phillippe Leite fr the purpse f this wrkshp Plicy questins are causal

More information