Department of Computer Science Artificial Intelligence Research Laboratory. Iowa State University MACHINE LEARNING

Size: px
Start display at page:

Download "Department of Computer Science Artificial Intelligence Research Laboratory. Iowa State University MACHINE LEARNING"

Transcription

1 MACHINE LEARNING Vasant Honavar Bonformatcs and Computatonal Bology Program Center for Computatonal Intellgence, Learnng, & Dscovery Iowa State Unversty 1

2 A Recpe for Learnng P ( h D) P( D h) P( h) P( D) P (h) pror probablty of hypothess h P (D) pror probablty of tranng data D P (h D) probablty of h gven D P (D h) probablty of D gven h

3 3 Iowa State Unversty A Recpe for learnng ) ( arg max ), ( ) (, If Maxmum lkelhood hypothess ) ( ) ( arg max ) ( ) ( ) ( arg max ) ( arg max hypothess posteror Maxmum a h D P h h P h P H h h h P h D P D P h P h D P D h P h H h ML H h H h H h MAP

4 Brute Force MAP Hypothess Learner For each hypothess h n H, calculate the posteror probablty P( D h) P( h) P ( h D) P( D) Output the hypothess wth the hghest posteror probablty h arg max P( D h ) P( h ) h MAP ML h H arg max P( D h h H ) 4

5 h MAP log - log arg mn ( h ) P( D h Brute Force MAP Hypothess Learner arg max P( D h arg max P h H h H h H ) P( h ( log P( D h ) + log P( h )) ( log P( D h ) log P( h )) ) descrpton length of under optmal Bayesan learnng mples Occam's razor! descrpton length of h of hypotheses ) under optmal encodng data D gven h encodng based on P( D h ) 5

6 MAP Learnng of -class classfers Consder a -class learnng problem defned over an Instance space X, hypothess space H, tranng examples D {e (X, d ) X X; d c(x )} Consder a learnng algorthm whch outputs a hypothess that s consstent wth the examples n D Let V H,D be the subset of hypotheses n H that are consstent wth D (the verson space) What would Bayes rule produce as the MAP hypothess? 6

7 MAP Learnng of Bnary Concepts (-class classfers) Assumptons 1. The hypothess space s fnte..the tranng examples are..d. 3.The tranng data s nose free. d h H, c P ( X ) 4.The target concept c H 5. d, ( h) All hypotheses are a pror equally lkely : 1 H 7

8 8 Iowa State Unversty MAP Learnng of -class classfers ( ) ( ) ( ) ( ) ( ) ( ) ( ) 0 otherwse ), ( ; ) ( f 1 ), ( ), (, are ndependent of Assumng,,. samples mply.. X h d P d X h X h d P X P X h d P h D P h X h X P X h d P h d X P h e P h D P d

9 MAP Learnng of -class classfers If {1... m}, d If h s consstent wth D, that s, P {1... m}such that h ( X ) ( D h) P( X ) 1 h s consstent wth D, that s, P( D h) d 0 m then h ( X ) then 9

10 10 Iowa State Unversty MAP Learnng of -class classfers ( ) ( ) ( ) ( ) ( ) ( ) ( ) ( ) 0 ) ( we have, and 1 ) ( we have, 0 1 ) ( ) (,,,,,, + D h P V h V X P V H H X P D P h P D h P D h P V h X P H V X P H h P h D P D P D H D H D H D H D H H h V h H h D H Every hypothess n H that s consstent wth D s a MAP hypothess

11 Bayesan Recpe for Learnng If the tranng examples are ndependently dentcally dstrbuted nose-free And f each canddate hypothess n H s equally a pror equally lkely Then every hypothess that s consstent wth the tranng data (that s, correctly classfes each tranng example) maxmzes P(h D) In ths settng, all that the learner has to do s to produce a hypothess from H that s consstent wth the tranng data 11

12 Effect of Data (Evdence) on the Verson Space 1

13 Bayesan Recpe for Learnng h MAP ( ) ( ) arg mn log P D h log P h h H The number of bts needed to encode the Data D gven the hypothess h (under optmal encodng) n other words, error of h or exceptons to h The number of bts needed to encode the hypothess h (under optmal encodng) 13

14 Bayesan Recpe for Learnng Suppose the tranng data are d and nose-free but each hypothess n H s not equally a pror equally lkely h MAP arg max P( D h ) P( h ) h H arg max h H h H arg mn ( log P( D h ) + log P( h )) ( log P( D h ) log P( h )) 14

15 Bayesan Recpe for Learnng If the tranng data are d and nose-free but each hypothess n H s not equally a pror equally lkely then the learner has to produce a hypothess from H that trades off the error on the tranng data aganst complexty of the hypothess Decson tree classfers, whch we wll look at next, provde an example of ths approach 15

16 Learnng Decson Tree Classfers On average, the nformaton needed to convey the class H P r membershp of a random nstance drawn from nature s ( ) Nature Instance Tranng Data S S 1 S S m Classfer Class label ) H 1 ) where P s an estmate of P m ( P) ( pˆ ) log ( pˆ ) Hˆ ( X ) and X s a random varable wth dstrbuton P S s the mult-set of examples belongng to class C 16

17 Learnng Decson Tree Classfers The task of the learner then s to extract the needed nformaton from the tranng set and store t n the form of a decson tree for classfcaton Informaton gan based decson tree learner Start wth the entre tranng set at the root Recursvely add nodes to the tree correspondng to tests that yeld the greatest expected reducton n entropy (or the largest expected nformaton gan) untl some termnaton crteron s met ( e.g., the tranng data at every leaf node has zero entropy ) 17

18 Learnng Decson Tree Classfers - Example Instances ordered 3-tuples of attrbute values correspondng to Heght (tall, short) Har (dark, blonde, red) Eye (blue, brown) Tranng Data Instance Class label I 1 (t, d, l) + I (s, d, l) + I 3 (t, b, l) I 4 (t, r, l) I 5 (s, b, l) I 6 (t, b, w) + I 7 (t, d, w) + I 8 (s, b, w) + 18

19 t Learnng Decson Tree Classfers - Example I 1 I Heght s Hˆ ( X ) log log bts Hˆ 3 3 ( X Heght t) log log bts S t I 1 I 3 I 4 I 6 I 7 S s I I 5 I Hˆ 5 ( X Heght) Hˆ ( X Heght t) + Hˆ ( X Heght s) ( ) + ( ) 0. 95bts Smlarly, Hˆ ( X Eye) bts Hˆ 1 1 ( X Heght s) log log bts and Hˆ ( X Har) 0. 5bts Har s the most nformatve because t yelds the largest reducton n entropy. Test on the value of Har s chosen to correspond to the root of the decson tree 19

20 Learnng Decson Tree Classfers - Example d Har + Eye - - l b w + r Compare the result wth Naïve Bayes In practce, we need some way to prune the tree to avod overfttng the tranng data more on ths later. 0

21 Learnng, generalzaton, overfttng Consder the error of a hypothess h over tranng data: Error Tran (h) entre dstrbuton D of data: Error D (h) Hypothess h H over fts tranng data f there s an alternatve hypothess h H such that and Error Tran ( h) < Error Tran ( h' ) Error D ( h) > Error D ( h' ) 1

22 Over fttng n decson tree learnng

23 Causes of over fttng As we move further away from the root, the data set used to choose the best test becomes smaller poor estmates of entropy Nosy examples can further exacerbate over fttng 3

24 Mnmzng over fttng Use roughly the same sze sample at every node to estmate entropy when there s a large data set from whch we can sample Stop when further splt fals to yeld statstcally sgnfcant nformaton gan (estmated from valdaton set) Grow full tree, then prune mnmze sze (tree) + sze (exceptons (tree)) 4

25 Reduced error prunng Each decson node n the tree s consdered as a canddate for prunng Prunng a decson node conssts of removng the sub tree rooted at that node, makng t a leaf node, and assgnng t the most common label at that node or storng the class dstrbuton at that node (for probablstc classfcaton) 5

26 Reduced error prunng Example A 100 Node A Accuracy gan by Prunng -0% Aa 1 Aa + B B +10% A Aa 1 Aa Before Prunng + - After Prunng 6

27 Reduced error prunng Do untl further prunng s harmful: Evaluate mpact on valdaton set of prunng each canddate node Greedly select a node whch most mproves the performance on the valdaton set when the sub tree rooted at that node s pruned Drawback holdng back the valdaton set lmts the amount of tranng data avalable not desrable when data set s small 7

28 Reduced error prunng Pruned Tree 8

29 Prunng Evaluate Canddate splt to decde f the resultng nformaton gan s sgnfcantly greater than zero as determned usng a sutable hypothess testng method at a desred sgnfcance level L R 50 n n n n 1 1L 1e 1e χ χ 50, n nl 50, nl 0, nl 50, p 0.5 N number of class1nstances (random splt) pn 1 5, n 5 ( n n ) ( n n ) 1L n 50, N n 1e 1e e + 1 pn L + n n e e Ths splt s sgnfcantly better than random wth confdence > 99% because χ >

30 Prunng based on whether nformaton gan s sgnfcantly greater than zero Evaluate Canddate splt to decde f the resultng nformaton gan s sgnfcantly greater than zero as determned usng a sutable hypothess testng method at a desred sgnfcance level χ Example: statstc In the -class, bnary (L,R) splt case, n 1 of class 1, n of class ; Nn 1 +n Splt sends pn to L and (1-p)N to R χ ( n ) L ne 1 Random splt would send pn 1 of class 1 to L and pn of class to L The crtcal value of χ depends on the degrees of freedom whch s 1 n ths case (for a gven p, n 1L fully specfes n L n 1R and n R ) In general, the number of degrees of freedom can be > 1 n e 30

31 Prunng based on whether nformaton gan s sgnfcantly greater than zero The greater the value of χ Branches 1Classes ( n ne ) the less lkely t s that the splt s random. For 1 1 n a suffcently hgh value χ N n p [ p n e n p p n... p +... n Classes Branches ]; e Branches 1 p 1 of χ, the dfference between the expected (random) splt s statstcally sgnfcant and we reect the null hypothess that the splt s random. Degrees of freedom ( Classes 1)( Branches 1) 31

32 Rule post-prunng Convert tree to equvalent set of rules IF THEN IF THEN... (Outlook Sunny) (Humdty Hgh) PlayTenns No (Outlook Sunny) (Humdty Hgh) PlayTenns Yes 3

33 Rule post-prunng 1. Convert tree to equvalent set of rules. Prune each rule ndependently of others by droppng a condton at a tme f dong so does not reduce estmated accuracy (at the desred confdence level) 3. Sort fnal rules n order of lowest to hghest error Advantage can potentally correct bad choces made close to the root Post prunng based on valdaton set s the most commonly used method n practce Development of pre prunng methods wth comparable performance that do not requre a valdaton set s an open problem 33

34 Classfcaton of nstances Unque classfcaton possble when each leaf has zero entropy and there are no mssng attrbute values Most lkely classfcaton based on dstrbuton of classes at a node when there are no mssng attrbute values Probablstc classfcaton based on dstrbuton of classes at a node when there are no mssng attrbute values 34

35 Handlng dfferent types of attrbute values Types of attrbutes Nomnal values are names Ordnal values are ordered Cardnal (Numerc) values are numbers (hence ordered). 35

36 Handlng numerc attrbutes Attrbute T Class N N Y Y Y N Canddate splts Sort nstances by value of numerc attrbute under consderaton For each attrbute, fnd the test whch yelds the lowest entropy Greedly choose the best test across all attrbutes T > ( )? T > ( )? E( S T > 49?) (0) + log log

37 Handlng numerc attrbutes Axs-parallel splt Oblque splt C C C 1 C 1 Oblque splts cannot be realzed by unvarate tests 37

38 Two-way versus mult-way splts Entropy crteron favors many-valued attrbutes Pathologcal behavor what f n a medcal dagnoss data set, socal securty number s one of the canddate attrbutes? A value versus A value Solutons Only two-way splts (CART) Entropy rato (C4.5) EntropyRat o( S SpltEntro py( S A) A) Entropy ( S A) SpltEntro py( S A) Values ( A) S S log 1 S S 38

39 Alternatve splt crtera Consder splt of set S based on attrbute A Impurty( S A) Values( A) 1 ImpurtyS ( ) Classes Z Z Entropy Impurty( Z) log 1 Z Z Classes Z Z Z Gn Impurty( Z) 1 Z Z 1 Z (Expected rate of error f class label s pcked randomly accordng to dstrbuton of nstances n a set) 39

40 Alternatve splt crtera One-sded splt crtera often useful for exploratory analyss of data Impurty( S A) Mn Values( A) { Impurty( S )} 40

41 Incorporatng Attrbute costs Not all attrbute measurements are equally costly or rsky In Medcal dagnoss Blood-Test has cost $150 Exploratory-Surgery may have a cost of $3000 Goal: Learn a Decson Tree Classfer whch mnmzes cost of classfcaton Tan and Schlmmer (1990) Nunez (1988) Gan( S, A) 1 ( Cost ( A) + 1) Gan ( S, A) Cost( A) where w [0, 1] determnes mportance of cost w 41

42 Incorporatng Dfferent Msclassfcaton Costs for dfferent classes Not all msclassfcatons are equally costly An occasonal false alarm about a nuclear power plant meltdown s less costly than the falure to alert when there s a chance of a meltdown Weghted Gn Impurty Impurty( S) S S λ S S λ s the cost of wrongly assgnng an nstance belongng to class to class 4

43 Dealng wth Mssng Attrbute Values (Soluton 1) Sometmes, the fact that an attrbute value s mssng mght tself be nformatve Mssng blood sugar level mght mply that the physcan had reason not to measure t Introduce a new value (one per attrbute) to denote a mssng value Decson tree constructon and use of tree for classfcaton proceed as before 43

44 Dealng wth Mssng Attrbute Values Soluton Durng decson tree constructon Replace a mssng attrbute value n a tranng example wth the most frequent value found among the nstances at the node Durng use of tree for classfcaton Replace a mssng attrbute value n an nstance to be classfed wth the most frequent value found among the tranng nstances at the node 44

45 Dealng wth Mssng Attrbute Values (Soluton 3) Durng decson tree constructon Replace a mssng attrbute value n a tranng example wth the most frequent value found among the nstances at the node that have the same class label as the tranng example Durng use of tree for classfcaton Assgn to a mssng attrbute the most frequent value found at the node (based on the tranng sample) Sort the nstance through the tree to generate the class label 45

46 Dealng wth Mssng Attrbute Values (Soluton 4) Durng decson tree constructon Generate several fractonally weghted tranng examples based on the dstrbuton of values for the correspondng attrbute at the node Durng use of tree for classfcaton Generate multple nstances by assgnng canddate values for the mssng attrbute based on the dstrbuton of nstances at the node Sort each such nstance through the tree to generate canddate labels and assgn the most probable class label or probablstcally assgn class label 46

47 Dealng wth Mssng Attrbute Values A 1 0 n + 60, n - 40 (n + A1)50 + B (n + A0) 10; (n - A0 ) (n - A0, B1) (n + A0, B0) 10 Suppose B s mssng Replacement wth most frequent value at the node B1 Replacement wth most frequent value f class s + B0 47

48 Dealng wth Mssng Attrbute Values A 1 0 n + 60, n - 40 Suppose B s mssng (n + A1)50 + B (n + A0) 10; (n - A0 ) (n (n - A0, B1)40 + A0, B0) Fractonal nstance based on dstrbuton at the node.. 4/5 for B1, 1/5 for B0 Fractonal nstance based on dstrbuton at the node for class +.. 1/5 for B0, 0 for B1 48

49 Recent Developments n Decson Tree Classfers Learnng Decson Trees from Dstrbuted Data (Caragea, Slvescu and Honavar, 004) Learnng Decson Trees from Attrbute Value Taxonomes and partally specfed data (Zhang and Honavar, 003; 004; Zhang et al., 005) Learnng Decson Trees from Relatonal Data (Atramentov, Leva, and Honavar, 003) Decson Trees for Mult-label Classfcaton Tasks Learnng Decson Trees from Very Large Data Sets Learnng Decson Trees from Data Streams 49

50 Summary of Decson Trees Smple Fast (Lnear n sze of the tree, lnear n the sze of the tranng set, lnear n the number of attrbutes) Produce easy to nterpret rules Good for generatng smple predctve rules from data wth lots of attrbutes 50

51 Neural Networks Decson trees are good at modelng nonlnear nteractons among a small subset of attrbutes Sometmes we are nterested n lnear nteractons among all attrbutes Smple neural networks are good at modelng such nteractons The resultng models have close connectons wth naïve Bayes. 51

52 Measurng classfer performance To smplfy matters, assume that class labels are bnary M-class problem s turned nto M -class problems 5

53 Measurng Classfer Performance N: Total number of nstances n the data set TP : Number of True postves for class FP : Number of False postves for class TN : Number of True Negatves for class FN : Number of False Negatves for class Accuracy TP + TN N P class ( c label c ) Perfect classfer Accuracy 1 Popular measure Based n favor of the maorty class! Should be used wth cauton! 53

54 Measurng Classfer Performance: Senstvty Senstvty TP TP ( label c class c ) Count( class c ) ( c class c ) Count + FN P label Perfect classfer Senstvty 1 Probablty of correctly labelng members of the target class Also called recall or ht rate 54

55 Measurng Classfer Performance: Specfcty Specfcty TP TP ( label c class c ) Count( label c ) ( c label c ) Count + FP P class Perfect classfer Specfcty 1 Also called precson Probablty that a postve predcton s correct 55

56 C 1 Classfer Learnng -- Measurng Performance Class Label C 1 C 1 TP 55 FP5 C 1 FN10 TN30 N TP + FN + TN + TP senstvty1 TP + FN TP specfcty1 TP + FP TP + TN accuracy1 N FP

57 Measurng Performance: Precson, Recall, and False Alarm Rate Precson Specfcty TP TP + FP Recall Senstvty TP TP + FN Perfect classfer Precson1 Perfect classfer Recall1 FalseAlarm TN FP ( label c class c ) Count ( label c ) ( c class c ) Count P label + FP Perfect classfer False Alarm Rate 0 57

58 Measurng Performance Correlaton Coeffcent ( TP TN ) ( FP FN ) CC TP + FN TP + FP TN + FP TN + FN ( )( )( )( ) J 1 CC 1 Perfect classfer CC 1, Random guessng CC0 Corresponds to the standard measure of correlaton between two random varables Label and Class estmated from labels L and the correspondng class values C for the specal case of bnary (0/1) valued labels and classes CC where class d D label ( label label)( class class) 1ff σ JLABEL σ the true class of JCLASS 1ff the classfer assgns d d s class c to class c 58

59 Beware of termnologcal confuson n the lterature! Some bonformatcs authors use accuracy ncorrectly to refer to recall.e. senstvty or precson.e. specfcty In medcal statstcs, specfcty sometmes refers to senstvty for the negatve class.e. TN TN + FP Some authors use false alarm rate to refer to the probablty that a postve predcton s ncorrect.e. When you wrte provde the formula n terms of TP, TN, FP, FN When you read check the formula n terms of TP, TN, FP, FN FP FP + TP 1 Precson 59

60 Measurng Classfer Performance TP, FP, TN, FN provde the relevant nformaton No sngle measure tells the whole story A classfer wth 90% accuracy can be useless f 90 percent of the populaton does not have cancer and the 10% that do are msclassfed by the classfer Use of multple measures recommended Beware of termnologcal confuson! 60

61 Measurng Classfer Performance Mcro-averages Mcro averagng gves equal mportance to each nstance Classes wth large number of nstances domnate TP TP McroAverage Precson McroAvera ge McroAvera ge CC TP + TP + FP TP TN FN TP + FP McroAverage Recall FalseAlarm 1 McroAverage Precson McroAverage Accuracy TP N Etc. FP FN TN + FP TP TN + + FN FN 61

62 Measurng Classfer Performance Macro-averagng Macro averagng gves equal mportance to each of the M classes MacroAverage Senstvty 1 M Senstvty MacroAverageCorrelatonCoeff 1 M CorrelatonCoeff MacroAverage Specfcty 1 M Specfcty 6

63 Recever Operatng Characterstc (ROC) Curve We can often trade off recall versus precson e.g., by adustng classfcaton threshold θ e.g., label c f P ( c X ) ( c X ) P > θ ROC curve s a plot of Senstvty aganst False Alarm Rate whch characterzes ths tradeoff for a gven classfer 63

64 Recever Operatng Characterstc (ROC) Curve 1 Perfect classfer TP Senstvty TP + FN False Alarm Rate FP FP + TN 64

65 Measurng Performance of Classfers ROC curves ROC curves offer a more complete pcture of the performance of the classfer as a functon of the classfcaton threshold A classfer h s better than another classfer g f ROC(h) domnates the ROC(g) ROC(h) domnates ROC(g) AreaROC(h) > AreaROC(g) 65

66 Evaluatng a Classfer How well can a classfer be expected to perform on novel data? We can estmate the performance (e.g., accuracy, senstvty) of the classfer usng a test data set (not used for tranng) How close s the estmated performance to the true performance? References: Evaluaton of dscrete valued hypotheses Chapter 5, Mtchell Emprcal Methods for Artfcal Intellgence, Cohen Better yet, take Stat 430x 66

67 Estmatng the performance of a classfer The true error of a hypothess h wth respect to a target functon f and an nstance dstrbuton D s [ f ( x) h( )] Error ( h) Pr x D x D The sample error of a hypothess h wth respect to a target functon f and an nstance dstrbuton D s Error S ( h) δ ( a, b) 1ff 1 S x S δ ( f ( x) h( x)) a b; δ ( a, b) 0 otherwse 67

68 68 Iowa State Unversty Estmatng classfer performance ( ) ( ) ( ) [ ] ) ( ) ( ) ( ) ( Pr,,, ) ( },,, { ) ( c X D a X D x f x h h error x f x h d c b a x X D d c b a X Doman D D

69 Evaluatng the performance of a classfer Sample error estmated from tranng data s an optmstc estmate Bas E Error ( h) Error ( h [ ] ) S D For an unbased estmate, h must be evaluated on an ndependent sample S (whch s not the case f S s the tranng set!) Even when the estmate s unbased, t can vary across samples! If h msclassfes 8 out of 100 samples 8 Error S ( h) 100 How close s the sample error to the true error?

70 How close s the estmated error to the true error? Choose a sample S of sze n accordng to dstrbuton D Measure Error S (h) (h) Error S s a random varable (outcome of a random experment) Gven Error S ( h), what can we conclude about Error D ( h)? More generally, gven the estmated performance of a hypothess, what can we say about ts actual performance? 70

71 Evaluaton of a classfer wth lmted data There s extensve lterature on how to estmate classfer performance from samples and how to assgn confdence to estmates (See Mtchell, Chapter 5) Holdout method use part of the data for tranng, and the rest for testng We may be unlucky tranng data or test data may not be representatve Soluton Run multple experments wth dsont tranng and test data sets n whch each class s represented n roughly the same proporton as n the entre data set 71

72 Estmatng the performance of the learned classfer K-fold cross-valdaton Partton the data (mult) set S nto K equal parts S 1..S K wth roughly the same class dstrbuton as S. Errorc 0 For 1 to K do Error { S α Test S STran S S ; Learn( S Tran ) Errorc Errorc + Error( α, STest ) Errorc K ; Output( Error) } 7

73 Leave-one-out cross-valdaton K-fold cross valdaton wth K n where n s the total number of samples avalable n experments usng n-1 samples for tranng and the remanng sample for testng Leave-one-out cross-valdaton does not guarantee the same class dstrbuton n tranng and test data! Extreme case: 50% class 1, 50% class Predct maorty class label n the tranng data True error 50%; Leave-one-out error estmate 100%!!!!! 73

74 Estmatng classfer performance Recommended procedure Use K-fold cross-valdaton (K5 or 10) for estmatng performance estmates (accuracy, precson, recall, ponts on ROC curve, etc.) and 95% confdence ntervals around the mean Compute mean values of performance estmates and standard devatons of performance estmates Report mean values of performance estmates and ther standard devatons or 95% confdence ntervals around the mean Be skeptcal repeat experments several tmes wth dfferent random splts of data nto K folds! 74

75 Evaluatng a hypothess or a learnng algorthm How well can the decson tree be expected to perform on novel data? We can estmate the performance (e.g., accuracy) of the decson tree usng a test data set (not used for tranng) How close s the estmated performance to the true performance? Reference: Evaluaton of dscrete valued hypotheses Chapter 5, Mtchell 75

76 Evaluatng performance when we can afford to test on a large ndependent test set The true error of a hypothess h wth respect to a target functon f and an nstance dstrbuton D s [ f ( x) h( )] Error ( h) Pr x D x D The sample error of a hypothess h wth respect to a target functon f and an nstance dstrbuton D s Error S ( h) δ ( a, b) 1ff 1 S x S δ ( f ( x) a b; δ ( a, b) h( x)) 0 otherwse 76

77 Evaluatng Classfer performance Bas E Error [ ( h) ] Error ( h) S Sample error estmated from tranng data s an optmstc estmate For an unbased estmate, h must be evaluated on an ndependent sample S (whch s not the case f S s the tranng set!) Even when the estmate s unbased, t can vary across samples! 8 If h msclassfes 8 out of 100 samples Error S ( h) How close s the sample error to the true error? D

78 How close s estmated error to ts true value? Choose a sample S of sze n accordng to dstrbuton D Measure Error S (h) (h) Error S s a random varable (outcome of a random experment) Gven Error S ( h), what can we conclude about Error D ( h)? More generally, gven the estmated performance of a hypothess, what can we say about ts actual performance? 78

79 How close s estmated accuracy to ts true value? Queston: How close s p (the true probablty) to pˆ? Ths problem s an nstance of a well-studed problem n statstcs the problem of estmatng the proporton of a populaton that exhbts some property, gven the observed proporton over a random sample of the populaton. In our case, the property of nterest s that h correctly (or ncorrectly) classfes a sample. Testng h on a sngle random sample x drawn accordng to D amounts to performng a random experment whch succeeds f h correctly classfes x and fals otherwse. 79

80 How close s estmated accuracy to ts true value? The output of a hypothess whose true error s p as a bnary random varable whch corresponds to the outcome of a Bernoull tral wth a success rate p (the probablty of correct predcton) The number of successes r observed n N trals s a random varable Y whch follows the Bnomal dstrbuton n! P( r) 1 r!( n r)! r n r p ( p ) 80

81 Error S (h) s a Random Varable Probablty of observng r msclassfed examples n a sample of sze n: n! P( r) 1 r!( n r)! r n r p ( p ) r 81

82 Recall basc statstcs Consder a random experment wth dscrete valued outcomes The expected value of the correspondng random M varable Y s The varance of Y s y 1, y,... y M E ( Y) y Pr( Y y ) 1 [( Y E[ ]) ] Var( Y ) E Y The standard devaton of Y s σ Y Var(Y ) 8

83 How close s estmated accuracy to ts true value? The mean of a Bernoull tral wth success rate p p Varance p (1-p) If N trals are taken from the same Bernoull process, the observed success rate pˆ has the same mean p and varance p ( 1 p ) For large N, the dstrbuton of dstrbuton N pˆ follows a Gaussan 83

84 Bnomal Probablty Dstrbuton n! P( r) 1 r!( n r)! r n r p ( p ) Probablty P(r) of r heads n n con flps, f p Pr(heads) Expected, or mean value of X, E[X], s E[ X ] P( ) np 0 Varance of X s Var( X ) E[( X E[ X ]) ] np( 1 p) Standard devaton of X, σ X, s X N σ E[( X E[ X ]) ] np( 1 p) 84

85 Estmators, Bas, Varance, Confdence Interval r ErrorS ( h) n Error ( h) p D σ σ p( 1 p n ) Error S ( h) ErrorD ( h)( 1 ErrorD ( h n )) Error S ( h) σ ErrorS ( h)( 1 ErrorS ( h n )) Error S ( h) An N% confdence nterval for some parameter p that s the nterval whch s expected wth probablty N% to contan p 85

86 Normal dstrbuton approxmates bnomal Error S (h) follows a Bnomal dstrbuton, wth mean μ ErrorS ( h ) Error D ( h) standard devaton σ Errors S ( h) Error D ( h)( 1 Errors n D ( h)) We can approxmate ths by a Normal dstrbuton wth the same mean and varance when np(1-p) 5 86

87 p Normal dstrbuton 1 x μ 1 ( ) σ ( x) e πσ The probablty that X wll fall n the nterval (a, b) s gven by p ( x) dx b a Expected, or mean value of X s gven by E[X] μ Varance of X s gven by Var(X) σ Standard devaton of X s gven by σ X σ 87

88 How close s the estmated accuracy to ts true value? Let the probablty that a Gaussan random varable X, wth zero mean, takes a value between z and z, Pr[-z X z] c Pr[ X z] z Pr[ X z] 5% mples there s a 5% chance that X les more than 1.65 standard devatons from the mean, or Pr [-1.65 X 1.65 ] 90%

89 How close s the estmated accuracy to ts true value? pˆ But does not have zero mean and unt varance so we normalze to get pˆ p Pr z < < z c p( 1 p) n 89

90 How close s the estmated accuracy to ts true value? To fnd confdence lmts: Gven a partcular confdence fgure c, use the table to fnd the z correspondng to the probablty ½ (1-c). Use lnear nterpolaton for values not n the table p pˆ + z n ± z 1 + pˆ n z n pˆ n + z 4n 90

91 How close s the estmated accuracy to ts true value? Example pˆ 0.75; n 1000; c 0.80; z 1.8 Then wth 80% confdence, we can say that the value of p les n the nterval [0.733,0.768] Note: the normal dstrbuton assumpton s vald only for large n (.e. np(1-p) 5 or n > 30) so estmates based on smaller values of n should be taken wth a generous dose of salt 91

92 Estmatng confdence ntervals 80% of area (probablty) les n μ ±1.8σ N% of area (probablty) les n μ ±z N σ 9

93 Confdence ntervals If S contans n examples, drawn ndependently of h and each other and n 30 or np(1-p) 5, Then Wth approxmately N% probablty, Error S (h) les n nterval Error D ( h) ± Z N Error equvalently, Error D (h) les n nterval D ( h)( 1 Errors n D ( h)) Errors whch s approxmately S Errors ( h) ± Z S N ( h) ± Z N Error D ( h)( 1 Error n D ( h)) ErrorS ( h)( 1 Errors n S ( h)) 93

94 One sded confdence ntervals What s the probablty that Error D (h) s at most U? Symmetry of Gaussan dstrbuton mples that confdence nterval wth 1001 ( α)% confdence wth lower bound L and upper bound U corresponds to a confdence nterval wth confdence 100( 1 α )% and wth upper bound U but no lower bound (or vce versa) 94

95 General approach to dervng confdence ntervals 1. Identfy the populaton parameter p to be estmated e.g., Error D (h ). Defne a sutable estmator W preferably unbased, mnmum varance 3. Determne the dstrbuton D W obeyed by W, and the mean and varance of W 4. Determne the confdence nterval by fndng the thresholds L and U such that N% of the mass of the probablty dstrbuton D Y falls wthn the nterval [L,U]. 95

96 Central Lmt Theorem Smplfes Confdence Interval Calculatons Consder a set of ndependent, dentcally dstrbuted random varables Y 1... Y n, all governed by an arbtrary probablty dstrbuton wth mean μ and fnte varance σ. Defne the sample mean, Y n 1 n Y 1 Central Lmt Theorem As n, the dstrbuton governng approaches Y a Normal dstrbuton, wth mean μ and varance σ /n 96

97 Evaluaton of a classfer wth lmted data Holdout method use part of the data for tranng, and the rest for testng We may be unlucky tranng data or test data may not be representatve Soluton Run multple experments wth dsont tranng and test data sets n whch each class s represented n roughly the same proporton as n the entre data set 97

98 Estmatng the performance of the learned classfer K-fold cross-valdaton Partton the data (mult) set S nto K equal parts S 1..S K where each part has roughly the same class dstrbuton as S. A 0 For 1 to K do STran S S ; { STest S α Learn( S Tran ) A A + Accuracy( α, STest ) } Accuracy A/K; Output (Accuracy) 98

99 K-fold cross-valdaton Recommended procedure for evaluatng classfers when data are lmted Use K-fold cross-valdaton (K5 or 10) Better stll, repeat K-fold cross-valdaton R tmes and average the results 99

100 Dfference n error between two hypotheses We wsh to estmate d ErrorD ( h1 ) ErrorD ( h ) Suppose h 1 has been tested on a sample S 1 of sze n 1 drawn accordng to D and h has been tested on a sample S of sze n drawn accordng to D ) An unbased estmator d ErrorS ( h1 ) Error ( ) 1 S h For large n 1 and large n the correspondng error estmates follow Normal dstrbuton Dfference of two Normal dstrbutons yelds a normal dstrbuton wth varance equal to the sum of the varances of the ndvdual dstrbutons 100

101 Dfference between errors of two hypotheses d dˆ Error D ( h1 ) ErrorD ( h ) ErrorsS ( h1 ) Errors ( h 1 S ) σ dˆ Errors S ( h )( 1 Error ( h )) Error ( h )( S1 1 S S n 1 + Error n ( h )) dˆ ± z N Error S ( h )( 1 Error ( h )) Error ( h )( S1 1 S S n 1 Error When S 1 S, the varance of s smaller and the confdence nterval correct but overly conservatve + dˆ n ( h )) 101

102 Hypothess testng Is one hypothess lkely to be better than another? What s the probablty that ErrorD ( h1 ) > ErrorD ( h )? Suppose Error ( h ). 30; Error ( h ) 0. 0; dˆ S1 1 S. What s the probablty that d > 0 gven that dˆ 0. 10? Pr( d 0 dˆ 0. 10) Pr( dˆ < μ ˆ ) > d 10

103 Hypothess testng Pr( d If n1 n 100, σ ˆ d > 0 dˆ 0. 10) Pr( dˆ < μ ˆ σ ˆ ) d d We accept the hypothess that ErrorD ( h1 ) > ErrorD ( h wth 95% confdence ) Equvalently, we reect the opposte hypothess the null hypothess at a (1-0.95) 0.05 level of sgnfcance 103

104 Comparng learnng algorthms L A and L B Whch learnng algorthm s better at learnng f? Unlmted data Run L A and L B on large tranng set S tran drawn accordng to D Test the resultng hypotheses on a large ndependent test set S Test drawn accordng to D Estmate Pr[ Error ( L ( S )) > Error ( L ( S ))] D A Tran D B Tran Usng Error S Test ( LA( STran) ) and ErrorS ( LB( STran)) Test 104

105 Comparng learnng algorthms L A and L B Estmate the expected value of the dfference n errors of L A and L B where expectaton s taken over tranng sets S Tran drawn accordng to D S E Tran D [ Error L ( S )) Error ( L ( S ))] D( A Tran D B Tran We have a lmted data set S drawn from an unknown D!! 105

106 Comparng learnng algorthms L A and L B Lmted data Pared t-test Run L A and L B on large tranng set S Tran drawn accordng to D Test the resultng hypotheses on a large ndependent test set S Test drawn accordng to D Estmate Pr[ Error ( L ( S )) > Error S D Test A Tran Error ( L D B ( S Tran ))] ( LA( STran) ) and ErrorS ( LB( STran)) Test usng 106

107 Comparng learnng algorthms L A and L B Partton S nto k dsont test sets T 1, T,..., T k of equal sze For from 1 to k do { S Test T ; S Tran S - T δ } Return Error δ S Test ( LA( STran )) ErrorS ( LB( STran )) k 1 k δ 1 Pared t-test Test 107

108 Comparng learnng algorthms L A and L B For large test sets, each has Normal dstrbuton δ has Normal dstrbuton f δ are ndependent Can we estmate confdence nterval for as before? δ are not exactly ndependent because of samplng from S as opposed to the dstrbuton D (but we wll pretend that they are) We don t know the standard devaton of ths dstrbuton. So we estmate t from sample..but when the estmated varance s used, the dstrbuton s no longer Normal unless K s large (whch typcally t s not) δ δ 108

109 Comparng learnng algorthms L A and L B Approxmate N% confdence nterval for S E Tran S s gven by where [ Error L ( S )) Error ( L ( S ))] δ ± t N, k 1ψ δ ψ δ D( A Tran D B Tran k 1 ( δ δ ) k( k 1) 1 s the estmate of standard devaton of the t dstrbuton governng Z N and t plays a role analogous to that of N, K 1. δ As K, tn, K 1 Z N and ψ σ δ δ 109

110 Performance evaluaton summary Rgorous statstcal evaluaton s extremely mportant n expermental computer scence n general and machne learnng n partcular How good s a learned hypothess? Is one hypothess better than another? Is one learnng algorthm better than another on a partcular learnng task? (No learnng algorthm outperforms all others on all tasks No free lunch theorem) Dfferent procedures for evaluaton are approprate under dfferent condtons (large versus lmted versus small sample) Important to know when to use whch evaluaton method and be aware of pathologcal behavor (tendency to grossly overestmate or underestmate the target value under specfc condtons 110

MDL-Based Unsupervised Attribute Ranking

MDL-Based Unsupervised Attribute Ranking MDL-Based Unsupervsed Attrbute Rankng Zdravko Markov Computer Scence Department Central Connectcut State Unversty New Brtan, CT 06050, USA http://www.cs.ccsu.edu/~markov/ markovz@ccsu.edu MDL-Based Unsupervsed

More information

Predictive Analytics : QM901.1x Prof U Dinesh Kumar, IIMB. All Rights Reserved, Indian Institute of Management Bangalore

Predictive Analytics : QM901.1x Prof U Dinesh Kumar, IIMB. All Rights Reserved, Indian Institute of Management Bangalore Sesson Outlne Introducton to classfcaton problems and dscrete choce models. Introducton to Logstcs Regresson. Logstc functon and Logt functon. Maxmum Lkelhood Estmator (MLE) for estmaton of LR parameters.

More information

Department of Computer Science Artificial Intelligence Research Laboratory. Iowa State University MACHINE LEARNING

Department of Computer Science Artificial Intelligence Research Laboratory. Iowa State University MACHINE LEARNING MACHINE LEANING Vasant Honavar Bonformatcs and Computatonal Bology rogram Center for Computatonal Intellgence, Learnng, & Dscovery Iowa State Unversty honavar@cs.astate.edu www.cs.astate.edu/~honavar/

More information

Evaluation for sets of classes

Evaluation for sets of classes Evaluaton for Tet Categorzaton Classfcaton accuracy: usual n ML, the proporton of correct decsons, Not approprate f the populaton rate of the class s low Precson, Recall and F 1 Better measures 21 Evaluaton

More information

Chapter 11: Simple Linear Regression and Correlation

Chapter 11: Simple Linear Regression and Correlation Chapter 11: Smple Lnear Regresson and Correlaton 11-1 Emprcal Models 11-2 Smple Lnear Regresson 11-3 Propertes of the Least Squares Estmators 11-4 Hypothess Test n Smple Lnear Regresson 11-4.1 Use of t-tests

More information

Generative classification models

Generative classification models CS 675 Intro to Machne Learnng Lecture Generatve classfcaton models Mlos Hauskrecht mlos@cs.ptt.edu 539 Sennott Square Data: D { d, d,.., dn} d, Classfcaton represents a dscrete class value Goal: learn

More information

CS 2750 Machine Learning. Lecture 5. Density estimation. CS 2750 Machine Learning. Announcements

CS 2750 Machine Learning. Lecture 5. Density estimation. CS 2750 Machine Learning. Announcements CS 750 Machne Learnng Lecture 5 Densty estmaton Mlos Hauskrecht mlos@cs.ptt.edu 539 Sennott Square CS 750 Machne Learnng Announcements Homework Due on Wednesday before the class Reports: hand n before

More information

1. Inference on Regression Parameters a. Finding Mean, s.d and covariance amongst estimates. 2. Confidence Intervals and Working Hotelling Bands

1. Inference on Regression Parameters a. Finding Mean, s.d and covariance amongst estimates. 2. Confidence Intervals and Working Hotelling Bands Content. Inference on Regresson Parameters a. Fndng Mean, s.d and covarance amongst estmates.. Confdence Intervals and Workng Hotellng Bands 3. Cochran s Theorem 4. General Lnear Testng 5. Measures of

More information

Ensemble Methods: Boosting

Ensemble Methods: Boosting Ensemble Methods: Boostng Ncholas Ruozz Unversty of Texas at Dallas Based on the sldes of Vbhav Gogate and Rob Schapre Last Tme Varance reducton va baggng Generate new tranng data sets by samplng wth replacement

More information

Evaluation of classifiers MLPs

Evaluation of classifiers MLPs Lecture Evaluaton of classfers MLPs Mlos Hausrecht mlos@cs.ptt.edu 539 Sennott Square Evaluaton For any data set e use to test the model e can buld a confuson matrx: Counts of examples th: class label

More information

Logistic Regression. CAP 5610: Machine Learning Instructor: Guo-Jun QI

Logistic Regression. CAP 5610: Machine Learning Instructor: Guo-Jun QI Logstc Regresson CAP 561: achne Learnng Instructor: Guo-Jun QI Bayes Classfer: A Generatve model odel the posteror dstrbuton P(Y X) Estmate class-condtonal dstrbuton P(X Y) for each Y Estmate pror dstrbuton

More information

Boostrapaggregating (Bagging)

Boostrapaggregating (Bagging) Boostrapaggregatng (Baggng) An ensemble meta-algorthm desgned to mprove the stablty and accuracy of machne learnng algorthms Can be used n both regresson and classfcaton Reduces varance and helps to avod

More information

Lecture 4 Hypothesis Testing

Lecture 4 Hypothesis Testing Lecture 4 Hypothess Testng We may wsh to test pror hypotheses about the coeffcents we estmate. We can use the estmates to test whether the data rejects our hypothess. An example mght be that we wsh to

More information

Kernel Methods and SVMs Extension

Kernel Methods and SVMs Extension Kernel Methods and SVMs Extenson The purpose of ths document s to revew materal covered n Machne Learnng 1 Supervsed Learnng regardng support vector machnes (SVMs). Ths document also provdes a general

More information

Homework Assignment 3 Due in class, Thursday October 15

Homework Assignment 3 Due in class, Thursday October 15 Homework Assgnment 3 Due n class, Thursday October 15 SDS 383C Statstcal Modelng I 1 Rdge regresson and Lasso 1. Get the Prostrate cancer data from http://statweb.stanford.edu/~tbs/elemstatlearn/ datasets/prostate.data.

More information

Composite Hypotheses testing

Composite Hypotheses testing Composte ypotheses testng In many hypothess testng problems there are many possble dstrbutons that can occur under each of the hypotheses. The output of the source s a set of parameters (ponts n a parameter

More information

2E Pattern Recognition Solutions to Introduction to Pattern Recognition, Chapter 2: Bayesian pattern classification

2E Pattern Recognition Solutions to Introduction to Pattern Recognition, Chapter 2: Bayesian pattern classification E395 - Pattern Recognton Solutons to Introducton to Pattern Recognton, Chapter : Bayesan pattern classfcaton Preface Ths document s a soluton manual for selected exercses from Introducton to Pattern Recognton

More information

Chapter 13: Multiple Regression

Chapter 13: Multiple Regression Chapter 13: Multple Regresson 13.1 Developng the multple-regresson Model The general model can be descrbed as: It smplfes for two ndependent varables: The sample ft parameter b 0, b 1, and b are used to

More information

x = , so that calculated

x = , so that calculated Stat 4, secton Sngle Factor ANOVA notes by Tm Plachowsk n chapter 8 we conducted hypothess tests n whch we compared a sngle sample s mean or proporton to some hypotheszed value Chapter 9 expanded ths to

More information

10-701/ Machine Learning, Fall 2005 Homework 3

10-701/ Machine Learning, Fall 2005 Homework 3 10-701/15-781 Machne Learnng, Fall 2005 Homework 3 Out: 10/20/05 Due: begnnng of the class 11/01/05 Instructons Contact questons-10701@autonlaborg for queston Problem 1 Regresson and Cross-valdaton [40

More information

Negative Binomial Regression

Negative Binomial Regression STATGRAPHICS Rev. 9/16/2013 Negatve Bnomal Regresson Summary... 1 Data Input... 3 Statstcal Model... 3 Analyss Summary... 4 Analyss Optons... 7 Plot of Ftted Model... 8 Observed Versus Predcted... 10 Predctons...

More information

Module 3 LOSSY IMAGE COMPRESSION SYSTEMS. Version 2 ECE IIT, Kharagpur

Module 3 LOSSY IMAGE COMPRESSION SYSTEMS. Version 2 ECE IIT, Kharagpur Module 3 LOSSY IMAGE COMPRESSION SYSTEMS Verson ECE IIT, Kharagpur Lesson 6 Theory of Quantzaton Verson ECE IIT, Kharagpur Instructonal Objectves At the end of ths lesson, the students should be able to:

More information

Bayesian Learning. Smart Home Health Analytics Spring Nirmalya Roy Department of Information Systems University of Maryland Baltimore County

Bayesian Learning. Smart Home Health Analytics Spring Nirmalya Roy Department of Information Systems University of Maryland Baltimore County Smart Home Health Analytcs Sprng 2018 Bayesan Learnng Nrmalya Roy Department of Informaton Systems Unversty of Maryland Baltmore ounty www.umbc.edu Bayesan Learnng ombnes pror knowledge wth evdence to

More information

4 Analysis of Variance (ANOVA) 5 ANOVA. 5.1 Introduction. 5.2 Fixed Effects ANOVA

4 Analysis of Variance (ANOVA) 5 ANOVA. 5.1 Introduction. 5.2 Fixed Effects ANOVA 4 Analyss of Varance (ANOVA) 5 ANOVA 51 Introducton ANOVA ANOVA s a way to estmate and test the means of multple populatons We wll start wth one-way ANOVA If the populatons ncluded n the study are selected

More information

Comparison of Regression Lines

Comparison of Regression Lines STATGRAPHICS Rev. 9/13/2013 Comparson of Regresson Lnes Summary... 1 Data Input... 3 Analyss Summary... 4 Plot of Ftted Model... 6 Condtonal Sums of Squares... 6 Analyss Optons... 7 Forecasts... 8 Confdence

More information

MLE and Bayesian Estimation. Jie Tang Department of Computer Science & Technology Tsinghua University 2012

MLE and Bayesian Estimation. Jie Tang Department of Computer Science & Technology Tsinghua University 2012 MLE and Bayesan Estmaton Je Tang Department of Computer Scence & Technology Tsnghua Unversty 01 1 Lnear Regresson? As the frst step, we need to decde how we re gong to represent the functon f. One example:

More information

Feature Selection: Part 1

Feature Selection: Part 1 CSE 546: Machne Learnng Lecture 5 Feature Selecton: Part 1 Instructor: Sham Kakade 1 Regresson n the hgh dmensonal settng How do we learn when the number of features d s greater than the sample sze n?

More information

CIS526: Machine Learning Lecture 3 (Sept 16, 2003) Linear Regression. Preparation help: Xiaoying Huang. x 1 θ 1 output... θ M x M

CIS526: Machine Learning Lecture 3 (Sept 16, 2003) Linear Regression. Preparation help: Xiaoying Huang. x 1 θ 1 output... θ M x M CIS56: achne Learnng Lecture 3 (Sept 6, 003) Preparaton help: Xaoyng Huang Lnear Regresson Lnear regresson can be represented by a functonal form: f(; θ) = θ 0 0 +θ + + θ = θ = 0 ote: 0 s a dummy attrbute

More information

Machine learning: Density estimation

Machine learning: Density estimation CS 70 Foundatons of AI Lecture 3 Machne learnng: ensty estmaton Mlos Hauskrecht mlos@cs.ptt.edu 539 Sennott Square ata: ensty estmaton {.. n} x a vector of attrbute values Objectve: estmate the model of

More information

SDMML HT MSc Problem Sheet 4

SDMML HT MSc Problem Sheet 4 SDMML HT 06 - MSc Problem Sheet 4. The recever operatng characterstc ROC curve plots the senstvty aganst the specfcty of a bnary classfer as the threshold for dscrmnaton s vared. Let the data space be

More information

Statistics for Economics & Business

Statistics for Economics & Business Statstcs for Economcs & Busness Smple Lnear Regresson Learnng Objectves In ths chapter, you learn: How to use regresson analyss to predct the value of a dependent varable based on an ndependent varable

More information

Statistics II Final Exam 26/6/18

Statistics II Final Exam 26/6/18 Statstcs II Fnal Exam 26/6/18 Academc Year 2017/18 Solutons Exam duraton: 2 h 30 mn 1. (3 ponts) A town hall s conductng a study to determne the amount of leftover food produced by the restaurants n the

More information

/ n ) are compared. The logic is: if the two

/ n ) are compared. The logic is: if the two STAT C141, Sprng 2005 Lecture 13 Two sample tests One sample tests: examples of goodness of ft tests, where we are testng whether our data supports predctons. Two sample tests: called as tests of ndependence

More information

Lecture Notes on Linear Regression

Lecture Notes on Linear Regression Lecture Notes on Lnear Regresson Feng L fl@sdueducn Shandong Unversty, Chna Lnear Regresson Problem In regresson problem, we am at predct a contnuous target value gven an nput feature vector We assume

More information

Motion Perception Under Uncertainty. Hongjing Lu Department of Psychology University of Hong Kong

Motion Perception Under Uncertainty. Hongjing Lu Department of Psychology University of Hong Kong Moton Percepton Under Uncertanty Hongjng Lu Department of Psychology Unversty of Hong Kong Outlne Uncertanty n moton stmulus Correspondence problem Qualtatve fttng usng deal observer models Based on sgnal

More information

Classification as a Regression Problem

Classification as a Regression Problem Target varable y C C, C,, ; Classfcaton as a Regresson Problem { }, 3 L C K To treat classfcaton as a regresson problem we should transform the target y nto numercal values; The choce of numercal class

More information

First Year Examination Department of Statistics, University of Florida

First Year Examination Department of Statistics, University of Florida Frst Year Examnaton Department of Statstcs, Unversty of Florda May 7, 010, 8:00 am - 1:00 noon Instructons: 1. You have four hours to answer questons n ths examnaton.. You must show your work to receve

More information

Department of Quantitative Methods & Information Systems. Time Series and Their Components QMIS 320. Chapter 6

Department of Quantitative Methods & Information Systems. Time Series and Their Components QMIS 320. Chapter 6 Department of Quanttatve Methods & Informaton Systems Tme Seres and Ther Components QMIS 30 Chapter 6 Fall 00 Dr. Mohammad Zanal These sldes were modfed from ther orgnal source for educatonal purpose only.

More information

Classification learning II

Classification learning II Lecture 8 Classfcaton learnng II Mlos Hauskrecht mlos@cs.ptt.edu 539 Sennott Square Logstc regresson model Defnes a lnear decson boundar Dscrmnant functons: g g g g here g z / e z f, g g - s a logstc functon

More information

Basic Business Statistics, 10/e

Basic Business Statistics, 10/e Chapter 13 13-1 Basc Busness Statstcs 11 th Edton Chapter 13 Smple Lnear Regresson Basc Busness Statstcs, 11e 009 Prentce-Hall, Inc. Chap 13-1 Learnng Objectves In ths chapter, you learn: How to use regresson

More information

Statistical pattern recognition

Statistical pattern recognition Statstcal pattern recognton Bayes theorem Problem: decdng f a patent has a partcular condton based on a partcular test However, the test s mperfect Someone wth the condton may go undetected (false negatve

More information

x i1 =1 for all i (the constant ).

x i1 =1 for all i (the constant ). Chapter 5 The Multple Regresson Model Consder an economc model where the dependent varable s a functon of K explanatory varables. The economc model has the form: y = f ( x,x,..., ) xk Approxmate ths by

More information

Here is the rationale: If X and y have a strong positive relationship to one another, then ( x x) will tend to be positive when ( y y)

Here is the rationale: If X and y have a strong positive relationship to one another, then ( x x) will tend to be positive when ( y y) Secton 1.5 Correlaton In the prevous sectons, we looked at regresson and the value r was a measurement of how much of the varaton n y can be attrbuted to the lnear relatonshp between y and x. In ths secton,

More information

January Examinations 2015

January Examinations 2015 24/5 Canddates Only January Examnatons 25 DO NOT OPEN THE QUESTION PAPER UNTIL INSTRUCTED TO DO SO BY THE CHIEF INVIGILATOR STUDENT CANDIDATE NO.. Department Module Code Module Ttle Exam Duraton (n words)

More information

Lecture 20: Hypothesis testing

Lecture 20: Hypothesis testing Lecture : Hpothess testng Much of statstcs nvolves hpothess testng compare a new nterestng hpothess, H (the Alternatve hpothess to the borng, old, well-known case, H (the Null Hpothess or, decde whether

More information

9.913 Pattern Recognition for Vision. Class IV Part I Bayesian Decision Theory Yuri Ivanov

9.913 Pattern Recognition for Vision. Class IV Part I Bayesian Decision Theory Yuri Ivanov 9.93 Class IV Part I Bayesan Decson Theory Yur Ivanov TOC Roadmap to Machne Learnng Bayesan Decson Makng Mnmum Error Rate Decsons Mnmum Rsk Decsons Mnmax Crteron Operatng Characterstcs Notaton x - scalar

More information

Economics 130. Lecture 4 Simple Linear Regression Continued

Economics 130. Lecture 4 Simple Linear Regression Continued Economcs 130 Lecture 4 Contnued Readngs for Week 4 Text, Chapter and 3. We contnue wth addressng our second ssue + add n how we evaluate these relatonshps: Where do we get data to do ths analyss? How do

More information

UNIVERSITY OF TORONTO Faculty of Arts and Science. December 2005 Examinations STA437H1F/STA1005HF. Duration - 3 hours

UNIVERSITY OF TORONTO Faculty of Arts and Science. December 2005 Examinations STA437H1F/STA1005HF. Duration - 3 hours UNIVERSITY OF TORONTO Faculty of Arts and Scence December 005 Examnatons STA47HF/STA005HF Duraton - hours AIDS ALLOWED: (to be suppled by the student) Non-programmable calculator One handwrtten 8.5'' x

More information

2016 Wiley. Study Session 2: Ethical and Professional Standards Application

2016 Wiley. Study Session 2: Ethical and Professional Standards Application 6 Wley Study Sesson : Ethcal and Professonal Standards Applcaton LESSON : CORRECTION ANALYSIS Readng 9: Correlaton and Regresson LOS 9a: Calculate and nterpret a sample covarance and a sample correlaton

More information

Linear Approximation with Regularization and Moving Least Squares

Linear Approximation with Regularization and Moving Least Squares Lnear Approxmaton wth Regularzaton and Movng Least Squares Igor Grešovn May 007 Revson 4.6 (Revson : March 004). 5 4 3 0.5 3 3.5 4 Contents: Lnear Fttng...4. Weghted Least Squares n Functon Approxmaton...

More information

3.1 Expectation of Functions of Several Random Variables. )' be a k-dimensional discrete or continuous random vector, with joint PMF p (, E X E X1 E X

3.1 Expectation of Functions of Several Random Variables. )' be a k-dimensional discrete or continuous random vector, with joint PMF p (, E X E X1 E X Statstcs 1: Probablty Theory II 37 3 EPECTATION OF SEVERAL RANDOM VARIABLES As n Probablty Theory I, the nterest n most stuatons les not on the actual dstrbuton of a random vector, but rather on a number

More information

U-Pb Geochronology Practical: Background

U-Pb Geochronology Practical: Background U-Pb Geochronology Practcal: Background Basc Concepts: accuracy: measure of the dfference between an expermental measurement and the true value precson: measure of the reproducblty of the expermental result

More information

Chapter 5 Multilevel Models

Chapter 5 Multilevel Models Chapter 5 Multlevel Models 5.1 Cross-sectonal multlevel models 5.1.1 Two-level models 5.1.2 Multple level models 5.1.3 Multple level modelng n other felds 5.2 Longtudnal multlevel models 5.2.1 Two-level

More information

STAT 511 FINAL EXAM NAME Spring 2001

STAT 511 FINAL EXAM NAME Spring 2001 STAT 5 FINAL EXAM NAME Sprng Instructons: Ths s a closed book exam. No notes or books are allowed. ou may use a calculator but you are not allowed to store notes or formulas n the calculator. Please wrte

More information

Lecture 12: Classification

Lecture 12: Classification Lecture : Classfcaton g Dscrmnant functons g The optmal Bayes classfer g Quadratc classfers g Eucldean and Mahalanobs metrcs g K Nearest Neghbor Classfers Intellgent Sensor Systems Rcardo Guterrez-Osuna

More information

Answers Problem Set 2 Chem 314A Williamsen Spring 2000

Answers Problem Set 2 Chem 314A Williamsen Spring 2000 Answers Problem Set Chem 314A Wllamsen Sprng 000 1) Gve me the followng crtcal values from the statstcal tables. a) z-statstc,-sded test, 99.7% confdence lmt ±3 b) t-statstc (Case I), 1-sded test, 95%

More information

Hydrological statistics. Hydrological statistics and extremes

Hydrological statistics. Hydrological statistics and extremes 5--0 Stochastc Hydrology Hydrologcal statstcs and extremes Marc F.P. Berkens Professor of Hydrology Faculty of Geoscences Hydrologcal statstcs Mostly concernes wth the statstcal analyss of hydrologcal

More information

ECONOMICS 351*-A Mid-Term Exam -- Fall Term 2000 Page 1 of 13 pages. QUEEN'S UNIVERSITY AT KINGSTON Department of Economics

ECONOMICS 351*-A Mid-Term Exam -- Fall Term 2000 Page 1 of 13 pages. QUEEN'S UNIVERSITY AT KINGSTON Department of Economics ECOOMICS 35*-A Md-Term Exam -- Fall Term 000 Page of 3 pages QUEE'S UIVERSITY AT KIGSTO Department of Economcs ECOOMICS 35* - Secton A Introductory Econometrcs Fall Term 000 MID-TERM EAM ASWERS MG Abbott

More information

Department of Statistics University of Toronto STA305H1S / 1004 HS Design and Analysis of Experiments Term Test - Winter Solution

Department of Statistics University of Toronto STA305H1S / 1004 HS Design and Analysis of Experiments Term Test - Winter Solution Department of Statstcs Unversty of Toronto STA35HS / HS Desgn and Analyss of Experments Term Test - Wnter - Soluton February, Last Name: Frst Name: Student Number: Instructons: Tme: hours. Ads: a non-programmable

More information

Dr. Shalabh Department of Mathematics and Statistics Indian Institute of Technology Kanpur

Dr. Shalabh Department of Mathematics and Statistics Indian Institute of Technology Kanpur Analyss of Varance and Desgn of Experment-I MODULE VIII LECTURE - 34 ANALYSIS OF VARIANCE IN RANDOM-EFFECTS MODEL AND MIXED-EFFECTS EFFECTS MODEL Dr Shalabh Department of Mathematcs and Statstcs Indan

More information

Correlation and Regression. Correlation 9.1. Correlation. Chapter 9

Correlation and Regression. Correlation 9.1. Correlation. Chapter 9 Chapter 9 Correlaton and Regresson 9. Correlaton Correlaton A correlaton s a relatonshp between two varables. The data can be represented b the ordered pars (, ) where s the ndependent (or eplanator) varable,

More information

See Book Chapter 11 2 nd Edition (Chapter 10 1 st Edition)

See Book Chapter 11 2 nd Edition (Chapter 10 1 st Edition) Count Data Models See Book Chapter 11 2 nd Edton (Chapter 10 1 st Edton) Count data consst of non-negatve nteger values Examples: number of drver route changes per week, the number of trp departure changes

More information

j) = 1 (note sigma notation) ii. Continuous random variable (e.g. Normal distribution) 1. density function: f ( x) 0 and f ( x) dx = 1

j) = 1 (note sigma notation) ii. Continuous random variable (e.g. Normal distribution) 1. density function: f ( x) 0 and f ( x) dx = 1 Random varables Measure of central tendences and varablty (means and varances) Jont densty functons and ndependence Measures of assocaton (covarance and correlaton) Interestng result Condtonal dstrbutons

More information

Statistics for Managers Using Microsoft Excel/SPSS Chapter 13 The Simple Linear Regression Model and Correlation

Statistics for Managers Using Microsoft Excel/SPSS Chapter 13 The Simple Linear Regression Model and Correlation Statstcs for Managers Usng Mcrosoft Excel/SPSS Chapter 13 The Smple Lnear Regresson Model and Correlaton 1999 Prentce-Hall, Inc. Chap. 13-1 Chapter Topcs Types of Regresson Models Determnng the Smple Lnear

More information

ANSWERS. Problem 1. and the moment generating function (mgf) by. defined for any real t. Use this to show that E( U) var( U)

ANSWERS. Problem 1. and the moment generating function (mgf) by. defined for any real t. Use this to show that E( U) var( U) Econ 413 Exam 13 H ANSWERS Settet er nndelt 9 deloppgaver, A,B,C, som alle anbefales å telle lkt for å gøre det ltt lettere å stå. Svar er gtt . Unfortunately, there s a prntng error n the hnt of

More information

Lecture 3: Probability Distributions

Lecture 3: Probability Distributions Lecture 3: Probablty Dstrbutons Random Varables Let us begn by defnng a sample space as a set of outcomes from an experment. We denote ths by S. A random varable s a functon whch maps outcomes nto the

More information

princeton univ. F 13 cos 521: Advanced Algorithm Design Lecture 3: Large deviations bounds and applications Lecturer: Sanjeev Arora

princeton univ. F 13 cos 521: Advanced Algorithm Design Lecture 3: Large deviations bounds and applications Lecturer: Sanjeev Arora prnceton unv. F 13 cos 521: Advanced Algorthm Desgn Lecture 3: Large devatons bounds and applcatons Lecturer: Sanjeev Arora Scrbe: Today s topc s devaton bounds: what s the probablty that a random varable

More information

STAT 3008 Applied Regression Analysis

STAT 3008 Applied Regression Analysis STAT 3008 Appled Regresson Analyss Tutoral : Smple Lnear Regresson LAI Chun He Department of Statstcs, The Chnese Unversty of Hong Kong 1 Model Assumpton To quantfy the relatonshp between two factors,

More information

The Gaussian classifier. Nuno Vasconcelos ECE Department, UCSD

The Gaussian classifier. Nuno Vasconcelos ECE Department, UCSD he Gaussan classfer Nuno Vasconcelos ECE Department, UCSD Bayesan decson theory recall that we have state of the world X observatons g decson functon L[g,y] loss of predctng y wth g Bayes decson rule s

More information

INF 5860 Machine learning for image classification. Lecture 3 : Image classification and regression part II Anne Solberg January 31, 2018

INF 5860 Machine learning for image classification. Lecture 3 : Image classification and regression part II Anne Solberg January 31, 2018 INF 5860 Machne learnng for mage classfcaton Lecture 3 : Image classfcaton and regresson part II Anne Solberg January 3, 08 Today s topcs Multclass logstc regresson and softma Regularzaton Image classfcaton

More information

Course 395: Machine Learning - Lectures

Course 395: Machine Learning - Lectures Course 395: Machne Learnng - Lectures Lecture 1-2: Concept Learnng (M. Pantc Lecture 3-4: Decson Trees & CC Intro (M. Pantc Lecture 5-6: Artfcal Neural Networks (S.Zaferou Lecture 7-8: Instance ased Learnng

More information

ANSWERS CHAPTER 9. TIO 9.2: If the values are the same, the difference is 0, therefore the null hypothesis cannot be rejected.

ANSWERS CHAPTER 9. TIO 9.2: If the values are the same, the difference is 0, therefore the null hypothesis cannot be rejected. ANSWERS CHAPTER 9 THINK IT OVER thnk t over TIO 9.: χ 2 k = ( f e ) = 0 e Breakng the equaton down: the test statstc for the ch-squared dstrbuton s equal to the sum over all categores of the expected frequency

More information

DO NOT OPEN THE QUESTION PAPER UNTIL INSTRUCTED TO DO SO BY THE CHIEF INVIGILATOR. Introductory Econometrics 1 hour 30 minutes

DO NOT OPEN THE QUESTION PAPER UNTIL INSTRUCTED TO DO SO BY THE CHIEF INVIGILATOR. Introductory Econometrics 1 hour 30 minutes 25/6 Canddates Only January Examnatons 26 Student Number: Desk Number:...... DO NOT OPEN THE QUESTION PAPER UNTIL INSTRUCTED TO DO SO BY THE CHIEF INVIGILATOR Department Module Code Module Ttle Exam Duraton

More information

Maximum Likelihood Estimation (MLE)

Maximum Likelihood Estimation (MLE) Maxmum Lkelhood Estmaton (MLE) Ken Kreutz-Delgado (Nuno Vasconcelos) ECE 175A Wnter 01 UCSD Statstcal Learnng Goal: Gven a relatonshp between a feature vector x and a vector y, and d data samples (x,y

More information

MIMA Group. Chapter 2 Bayesian Decision Theory. School of Computer Science and Technology, Shandong University. Xin-Shun SDU

MIMA Group. Chapter 2 Bayesian Decision Theory. School of Computer Science and Technology, Shandong University. Xin-Shun SDU Group M D L M Chapter Bayesan Decson heory Xn-Shun Xu @ SDU School of Computer Scence and echnology, Shandong Unversty Bayesan Decson heory Bayesan decson theory s a statstcal approach to data mnng/pattern

More information

A Bayes Algorithm for the Multitask Pattern Recognition Problem Direct Approach

A Bayes Algorithm for the Multitask Pattern Recognition Problem Direct Approach A Bayes Algorthm for the Multtask Pattern Recognton Problem Drect Approach Edward Puchala Wroclaw Unversty of Technology, Char of Systems and Computer etworks, Wybrzeze Wyspanskego 7, 50-370 Wroclaw, Poland

More information

Expected Value and Variance

Expected Value and Variance MATH 38 Expected Value and Varance Dr. Neal, WKU We now shall dscuss how to fnd the average and standard devaton of a random varable X. Expected Value Defnton. The expected value (or average value, or

More information

Generalized Linear Methods

Generalized Linear Methods Generalzed Lnear Methods 1 Introducton In the Ensemble Methods the general dea s that usng a combnaton of several weak learner one could make a better learner. More formally, assume that we have a set

More information

BOOTSTRAP METHOD FOR TESTING OF EQUALITY OF SEVERAL MEANS. M. Krishna Reddy, B. Naveen Kumar and Y. Ramu

BOOTSTRAP METHOD FOR TESTING OF EQUALITY OF SEVERAL MEANS. M. Krishna Reddy, B. Naveen Kumar and Y. Ramu BOOTSTRAP METHOD FOR TESTING OF EQUALITY OF SEVERAL MEANS M. Krshna Reddy, B. Naveen Kumar and Y. Ramu Department of Statstcs, Osmana Unversty, Hyderabad -500 007, Inda. nanbyrozu@gmal.com, ramu0@gmal.com

More information

7.1. Single classification analysis of variance (ANOVA) Why not use multiple 2-sample 2. When to use ANOVA

7.1. Single classification analysis of variance (ANOVA) Why not use multiple 2-sample 2. When to use ANOVA Sngle classfcaton analyss of varance (ANOVA) When to use ANOVA ANOVA models and parttonng sums of squares ANOVA: hypothess testng ANOVA: assumptons A non-parametrc alternatve: Kruskal-Walls ANOVA Power

More information

Econ107 Applied Econometrics Topic 3: Classical Model (Studenmund, Chapter 4)

Econ107 Applied Econometrics Topic 3: Classical Model (Studenmund, Chapter 4) I. Classcal Assumptons Econ7 Appled Econometrcs Topc 3: Classcal Model (Studenmund, Chapter 4) We have defned OLS and studed some algebrac propertes of OLS. In ths topc we wll study statstcal propertes

More information

Lecture 6: Introduction to Linear Regression

Lecture 6: Introduction to Linear Regression Lecture 6: Introducton to Lnear Regresson An Manchakul amancha@jhsph.edu 24 Aprl 27 Lnear regresson: man dea Lnear regresson can be used to study an outcome as a lnear functon of a predctor Example: 6

More information

The Multiple Classical Linear Regression Model (CLRM): Specification and Assumptions. 1. Introduction

The Multiple Classical Linear Regression Model (CLRM): Specification and Assumptions. 1. Introduction ECONOMICS 5* -- NOTE (Summary) ECON 5* -- NOTE The Multple Classcal Lnear Regresson Model (CLRM): Specfcaton and Assumptons. Introducton CLRM stands for the Classcal Lnear Regresson Model. The CLRM s also

More information

Interval Estimation in the Classical Normal Linear Regression Model. 1. Introduction

Interval Estimation in the Classical Normal Linear Regression Model. 1. Introduction ECONOMICS 35* -- NOTE 7 ECON 35* -- NOTE 7 Interval Estmaton n the Classcal Normal Lnear Regresson Model Ths note outlnes the basc elements of nterval estmaton n the Classcal Normal Lnear Regresson Model

More information

where I = (n x n) diagonal identity matrix with diagonal elements = 1 and off-diagonal elements = 0; and σ 2 e = variance of (Y X).

where I = (n x n) diagonal identity matrix with diagonal elements = 1 and off-diagonal elements = 0; and σ 2 e = variance of (Y X). 11.4.1 Estmaton of Multple Regresson Coeffcents In multple lnear regresson, we essentally solve n equatons for the p unnown parameters. hus n must e equal to or greater than p and n practce n should e

More information

A Robust Method for Calculating the Correlation Coefficient

A Robust Method for Calculating the Correlation Coefficient A Robust Method for Calculatng the Correlaton Coeffcent E.B. Nven and C. V. Deutsch Relatonshps between prmary and secondary data are frequently quantfed usng the correlaton coeffcent; however, the tradtonal

More information

Joint Statistical Meetings - Biopharmaceutical Section

Joint Statistical Meetings - Biopharmaceutical Section Iteratve Ch-Square Test for Equvalence of Multple Treatment Groups Te-Hua Ng*, U.S. Food and Drug Admnstraton 1401 Rockvlle Pke, #200S, HFM-217, Rockvlle, MD 20852-1448 Key Words: Equvalence Testng; Actve

More information

Basic Statistical Analysis and Yield Calculations

Basic Statistical Analysis and Yield Calculations October 17, 007 Basc Statstcal Analyss and Yeld Calculatons Dr. José Ernesto Rayas Sánchez 1 Outlne Sources of desgn-performance uncertanty Desgn and development processes Desgn for manufacturablty A general

More information

Introduction to Regression

Introduction to Regression Introducton to Regresson Dr Tom Ilvento Department of Food and Resource Economcs Overvew The last part of the course wll focus on Regresson Analyss Ths s one of the more powerful statstcal technques Provdes

More information

Dr. Shalabh Department of Mathematics and Statistics Indian Institute of Technology Kanpur

Dr. Shalabh Department of Mathematics and Statistics Indian Institute of Technology Kanpur Analyss of Varance and Desgn of Experment-I MODULE VII LECTURE - 3 ANALYSIS OF COVARIANCE Dr Shalabh Department of Mathematcs and Statstcs Indan Insttute of Technology Kanpur Any scentfc experment s performed

More information

Statistical Inference. 2.3 Summary Statistics Measures of Center and Spread. parameters ( population characteristics )

Statistical Inference. 2.3 Summary Statistics Measures of Center and Spread. parameters ( population characteristics ) Ismor Fscher, 8//008 Stat 54 / -8.3 Summary Statstcs Measures of Center and Spread Dstrbuton of dscrete contnuous POPULATION Random Varable, numercal True center =??? True spread =???? parameters ( populaton

More information

Global Sensitivity. Tuesday 20 th February, 2018

Global Sensitivity. Tuesday 20 th February, 2018 Global Senstvty Tuesday 2 th February, 28 ) Local Senstvty Most senstvty analyses [] are based on local estmates of senstvty, typcally by expandng the response n a Taylor seres about some specfc values

More information

P R. Lecture 4. Theory and Applications of Pattern Recognition. Dept. of Electrical and Computer Engineering /

P R. Lecture 4. Theory and Applications of Pattern Recognition. Dept. of Electrical and Computer Engineering / Theory and Applcatons of Pattern Recognton 003, Rob Polkar, Rowan Unversty, Glassboro, NJ Lecture 4 Bayes Classfcaton Rule Dept. of Electrcal and Computer Engneerng 0909.40.0 / 0909.504.04 Theory & Applcatons

More information

8/25/17. Data Modeling. Data Modeling. Data Modeling. Patrice Koehl Department of Biological Sciences National University of Singapore

8/25/17. Data Modeling. Data Modeling. Data Modeling. Patrice Koehl Department of Biological Sciences National University of Singapore 8/5/17 Data Modelng Patrce Koehl Department of Bologcal Scences atonal Unversty of Sngapore http://www.cs.ucdavs.edu/~koehl/teachng/bl59 koehl@cs.ucdavs.edu Data Modelng Ø Data Modelng: least squares Ø

More information

Probability Theory (revisited)

Probability Theory (revisited) Probablty Theory (revsted) Summary Probablty v.s. plausblty Random varables Smulaton of Random Experments Challenge The alarm of a shop rang. Soon afterwards, a man was seen runnng n the street, persecuted

More information

Cluster Validation Determining Number of Clusters. Umut ORHAN, PhD.

Cluster Validation Determining Number of Clusters. Umut ORHAN, PhD. Cluster Analyss Cluster Valdaton Determnng Number of Clusters 1 Cluster Valdaton The procedure of evaluatng the results of a clusterng algorthm s known under the term cluster valdty. How do we evaluate

More information

VQ widely used in coding speech, image, and video

VQ widely used in coding speech, image, and video at Scalar quantzers are specal cases of vector quantzers (VQ): they are constraned to look at one sample at a tme (memoryless) VQ does not have such constrant better RD perfomance expected Source codng

More information

Probability and Random Variable Primer

Probability and Random Variable Primer B. Maddah ENMG 622 Smulaton 2/22/ Probablty and Random Varable Prmer Sample space and Events Suppose that an eperment wth an uncertan outcome s performed (e.g., rollng a de). Whle the outcome of the eperment

More information

Resource Allocation and Decision Analysis (ECON 8010) Spring 2014 Foundations of Regression Analysis

Resource Allocation and Decision Analysis (ECON 8010) Spring 2014 Foundations of Regression Analysis Resource Allocaton and Decson Analss (ECON 800) Sprng 04 Foundatons of Regresson Analss Readng: Regresson Analss (ECON 800 Coursepak, Page 3) Defntons and Concepts: Regresson Analss statstcal technques

More information

Linear Correlation. Many research issues are pursued with nonexperimental studies that seek to establish relationships among 2 or more variables

Linear Correlation. Many research issues are pursued with nonexperimental studies that seek to establish relationships among 2 or more variables Lnear Correlaton Many research ssues are pursued wth nonexpermental studes that seek to establsh relatonshps among or more varables E.g., correlates of ntellgence; relaton between SAT and GPA; relaton

More information