Resampling Methods. Chapter 5. Chapter 5 1 / 52

Size: px
Start display at page:

Download "Resampling Methods. Chapter 5. Chapter 5 1 / 52"

Transcription

1 Resampling Methds Chapter 5 Chapter 5 1 / 52

2 1 51 Validatin set apprach 2 52 Crss validatin 3 53 Btstrap Chapter 5 2 / 52

3 Abut Resampling An imprtant statistical tl Pretending the data as ppulatin and repeatedly draw sample frm the data Main task: assess the validity/accuracy f statistical methds and mdels Of particular imprtance fr statistical predictin, by assessing test errr Crss-validatin and btstrap are tw types t be addressed in this chapter Chapter 5 3 / 52

4 Validatin and Crss validatin Validatin set apprach LOOCV (Leave-ne-ut crss valiatin) K-fld crss validatin Chapter 5 4 / 52

5 Btstrap Sampling with replacement (typically) n times, n is the sample size f data Especially useful in statistical inference (can be even mre accurate than nrmal apprximatin) All purpse resampling prcedure Used in, fr example, bagging and randm frest Chapter 5 5 / 52

6 51 Validatin set apprach traing errr easily cmputable with training data because f pssibility f ver-fit, it cannt be used t prperly assess test errr It is pssible t estimate the test errr, by, fr example, making adjustments f the training errr The adjusted R-squared, AIC, BIC, etc serve this purpse These methds rely n certain assumptins and are nt fr general purpse Chapter 5 6 / 52

7 51 Validatin set apprach Test errr wuld be als easily cmputable, if test data are well designated Nrmally we are just given data Shall have t create test data fr the purpse f cmputing test errr Artifically separate data int traning data and test data fr validatin purpse is called crss-validatin The test data here shuld be mre accurately called validatin data r hld ut data, meaning that they nt used in training mdels Mdel fitting nly uses the training data Chapter 5 7 / 52

8 51 Validatin set apprach In a data-rich scenari, we can affrd t separate the data int three parts training data: used t train varius mdels validatin data: used t assess the mdels and identify the best test data: test the results f the best mdel Usually, peple als call validatin data r hld-ut data as test data Chapter 5 8 / 52

9 51 Validatin set apprach Validatin set apprach Figure: 51 A schematic display f the validatin set apprach A set f n bservatins are randmly split int a training set (shwn in blue, cntaining bservatins 7, 22, and 13, amng thers) and a validatin set (shwn in beige, and cntaining bservatin 91, amng thers) The statistical learning methd is fit n the training set, and its perfrmance is evaluated n the validatin set Chapter 5 9 / 52

10 51 Validatin set apprach Example: Aut Data A nn-linear relatinship between mpg and hrsepwer mpg hrsepwer + hrsepwer 2 is better than mpg hrsepwer Shuld we add higher terms int the mdel?, like as cubic r even higher? One can check the p-values f regressin ceffeicients t answer the questin In fact, a mdel selectin prblem, and we can use validatin set apprach Chapter 5 10 / 52

11 51 Validatin set apprach Example: Aut Data randmly split the 392 bservatins int tw sets: a training set cntaining 196 f the data pints, and a validatin set cntaining the remaining 196 bservatins fit varius regressin mdels n the training sample The validatin set errr rates result frm evaluating their perfrmance n the validatin sample Here we MSE as a measure f validatin set errr, are shwn in the left-hand panel f Figure 52 Chapter 5 11 / 52

12 51 Validatin set apprach Validatin set apprach Mean Squared Errr Mean Squared Errr Degree f Plynmial Degree f Plynmial Figure: 52 The validatin set apprach was used n the Aut data set in rder t estimate the test errr that results frm predicting mpg using plynmial functins f hrsepwer Left: Validatin errr estimates fr a single split int training and validatin data sets Right: The validatin methd was repeated ten times, each time using a different randm split f the bservatins int a training set and a validatin set This illustrates the Chapter 5 12 / 52 variability in the estimated test MSE that results frm this apprach

13 51 Validatin set apprach Example: Aut Data The validatin set MSE fr the quadratic fit is cnsiderably smaller than fr the linear fit validatin set MSE fr the cubic fit is actually slightly larger than fr the quadratic fit This implies that including a cubic term in the regressin des NOT lead t better predictin than simply using a quadratic term Repeat the prcess f randmly splitting the sample set int tw parts, we will get a smewhat different estimate fr the test MSE Chapter 5 13 / 52

14 51 Validatin set apprach Example: Aut Data A quadratic term has a dramatically smaller validatin set MSE than the mdel with nly a linear term Nt much benefit in including cubic r higher-rder plynmial terms in the mdel Each f the ten curves results in a different test MSE estimate fr each f the ten regressin mdels cnsidered N cnsensus amng the curves as t which mdel results in the smallest validatin set MSE Based n the variability amng these curves, all that we can cnclude with any cnfidence is that the linear fit is nt adequate fr this data The validatin set apprach is cnceptually simple and is easy t i Chapter 5 14 / 52

15 51 Validatin set apprach A summary The validatin estimate f the test errr rate can be highly variable, depending n the randm split Only a subset f the bservatins the training set are used t fit the mdel Statistical methds tend t perfrm wrse when trained n fewer bservatins The validatin set errr rate may tend t verestimate the test errr rate fr the mdel fit n the entire data set Chapter 5 15 / 52

16 52 Crss validatin Crss validatin: vercme the drawback f validatin set apprach Our ultimate gal is t prduce the best mdel with best predictin accuracy Validatin set apprach has a drawback f using ONLY training data t fit mdel The validatin data d nt participate in mdel building but nly mdel assessment A waste f data We need mre data t participate in mdel building Chapter 5 16 / 52

17 52 Crss validatin Anther drawback f validatin set apprach It may ver-estimate the test errr fr the mdel with all data used t fit Statistical methds tend t perfrm wrse when trained n fewer bservatins The validatin set errr rate may tend t verestimate the test errr rate fr the mdel fit n the entire data set The methd f crss validatin culd vercme these drawbacks, by effectively using EVERY data pint in mdel building! Chapter 5 17 / 52

18 52 Crss validatin The leave-ne-ut crss-validatin Suppse the data cntain n data pints First, pick data pint 1 as validatin set, the rest as training set fit the mdel n the training set, evaluate the test errr, n the validatin set, dented as say MSE 1 Secnd, pick data pint 2 as validatin set, the rest as training set fit the mdel n the training set, evaluate the test errr n the validatin set, dented as say MSE 2 (repeat the prcedure fr all data pint) Obtain an estimate f the test errr by cmbining the MSE i, i =,, n: CV (n) = 1 n MSE i n i=1 Chapter 5 18 / 52

19 52 Crss validatin LOOCV Figure: 53 A schematic display f LOOCV A set f n data pints is repeatedly split int a training set (shwn in blue) cntaining all but ne bservatin, and a validatin set that cntains nly that bservatin (shwn in beige) The test errr is then estimated by averaging the n resulting MSE s The first training set cntains all but bservatin 1, the secnd training set cntains all but bservatin 2, and s frth Chapter 5 19 / 52

20 52 Crss validatin Advantages f LOOCV Far less bias, since the training data size (n 1) is clse t the entire data size (n) One single test errr estimate (thanks t the averaging), withut the variablity validatin set apprach A disadvantage: culd be cmputatinally expensive since the mdel need t be fit n times The MSE i may be t much crrelated Chapter 5 20 / 52

21 52 Crss validatin LOOCV applied t Aut data: LOOCV 10 fld CV Mean Squared Errr Mean Squared Errr Degree f Plynmial Degree f Plynmial Figure: 54 Crss-validatin was used n the Aut data set in rder t estimate the test errr that results frm predicting mpg using plynmial functins f hrsepwer Left: The LOOCV errr curve Right: 10-fld CV was run nine separate times, each with a different randm split f the data int ten parts The figure shws the nine slightly different CV errr curves Chapter 5 21 / 52

22 52 Crss validatin Cmplexity f LOOCV in linear mdel? Cnsinder linear mdel: y i = x T i β + ϵ i, i = 1,, n and the fitted values y i = x T i ˆβ, where ˆβ is the least squares estimate f β based n all data (x i, y i ), i = 1,, n Using LOOCV, the CV (n) = 1 n n i=1 (y i ŷ (i) i ) 2 where ŷ (i) i = xi T ˆβ (i) is the mdel predictr f y i based n the linear mdel fitted by all data except (x i, y i ) (delete ne), ie, ˆβ (i) is the least squares estimate f β based n all data but (x i, y i ) Chapter 5 22 / 52

23 52 Crss validatin Cmplexity f LOOCV in linear mdel? Lks t be cmplicated t cmpute least squares estimate n times Easy frmula: CV (n) = 1 n n i=1 ( yi ŷ i 1 h i ) 2 where ŷ i is the fitted values f least squares methd based n all data h i is the leverage Chapter 5 23 / 52

24 52 Crss validatin Simplicity f LOOCV in linear mdel One fit (with all data) des it all! The predictin errr rate (in terms f MSE) is just weighted average f the least squares fit residuals High leverage pint gets mre weight in predictin errr estimatin Chapter 5 24 / 52

25 52 Crss validatin K-fld crss validatin Divide the data int K subsets, usually f equal r similar sizes (n/k) Treat ne subset as validatin set, the rest tgether as a training set Run the mdel fitting n training set Calculate the test errr estimate n the validatin set, dented as MSE i, say Repeat the prcedures ver every subset Average ver the abve K estimates f the test errrs, and btain CV (K) = 1 K K MSE i i=1 LOOCV is a special case f K-fld crss validatin, actually n-fld crss validatin Chapter 5 25 / 52

26 52 Crss validatin K-fld crss validatin Figure: 55 A schematic display f 5-fld CV A set f n bservatins is randmly split int five nn-verlapping grups Each f these fifths acts as a validatin set (shwn in beige), and the remainder as a training set (shwn in blue) The test errr is estimated by averaging the five resulting MSE estimates Chapter 5 26 / 52

27 52 Crss validatin K-fld crss validatin Cmmn chices f K: K = 5 r K = 10 Advantage ver LOOCV: 1 cmputatinally lighter, espeically fr cmplex mdel with large data 2 Likely less variance (t be addressed later) Advantage ver validatin set apprach: Less variability resulting frm the data-split, thanks t the averaging Chapter 5 27 / 52

28 52 Crss validatin Mean Squared Errr Mean Squared Errr Mean Squared Errr Flexibility Flexibility Flexibility Figure: 56 True and estimated test MSE fr the simulated data sets in Figures 29 (left), 210 ( center), and 211 (right) The true test MSE is shwn in blue, the LOOCV estimate is shwn as a black dashed line, and the 10-fld CV estimate is shwn in range The crsses indicate the minimum f each f the MSE curves Chapter 5 28 / 52

29 52 Crss validatin Figure 29 Y Mean Squared Errr X Flexibility Figure: 29 Left: Data simulated frm f, shwn in black Three estimates f f are shwn: the linear regressin line (range curve), and tw smthing spline fits (blue and green curves) Right: Training MSE (grey curve), test MSE (red curve), and minimum pssible test MSE ver all methds (dashed line) Squares represent the training and test MSEs fr the three fits shwn in the left-hand panel Chapter 5 29 / 52

30 52 Crss validatin Figure 210 Y Mean Squared Errr X Flexibility Figure: 210 Details are as in Figure 29, using a different true f that is much clser t linear In this setting, linear regressin prvides a very gd fit t the data Chapter 5 30 / 52

31 52 Crss validatin Figure 211 Y Mean Squared Errr X Flexibility Figure: 211 Details are as in Figure 29, using a different f that is far frm linear In this setting, linear regressin prvides a very pr fit t the data Chapter 5 31 / 52

32 52 Crss validatin Special interests in the cmplexity parameter at miminum test errr A family f mdels indexed by a parameter, usually represeting flexibity r cmplexity f mdels Such parameter is ften called tuning parameter; it culd even be a number f variables Example: Order f plynmials f hrsepwer in the Aut data example Example: Penalizatin parameters in ridge, lass, etc (t be addressed in the next chapter) Intend t find the best mdel within this family, ie, t find the value f this tuning parameter Care less f the actual value f the test errr In the abve simulated data, all f the CV curves cme clse t identifying the crrect level f flexibility Chapter 5 32 / 52

33 52 Crss validatin Bias variance trade-ff In terms f bias f estimatin f test err: validatin set apprach has mre bias due t smaller size f training data; LOOCV is nearly unbiased; K-fld (eg, K=5 r 10) has intermediate bias In view f bias, LOOCV is mst preferred; and K-fld crss validatin next But, K-fld crss validatin has smaller variance than that f LOOCV The n traing sets LOOCV are t similar t each ther As a result, the trained mdels are t pstively crrelated The K training sets f K-fld crss validatin are much less similar t each ther As a result, the K-fld crss validatin generally has less variance than LOOCV Chapter 5 33 / 52

34 52 Crss validatin Crss validatin fr classificatin MSE is a ppular criteri t measure preditin/estimatin accuracy fr regressin There are ther criteria Fr classificatin with qualitative respnse, a natural chice is: 1 fr incrrect classificatin and 0 fr crrect classificatin Fr LOOCV, this leads t Err i = I(y i ŷ (i) i ), where y (i) i is the classificatin f i-th bservatin based n mdel fitted nt using i-th bservatin Then, CV (n) = 1 n n Err i which is just the average number f incrrect classifacitn i=1 Chapter 5 34 / 52

35 52 Crss validatin Example X 1 X2 Chapter 5 35 / 52

36 52 Crss validatin Figure 213 A simulated data set cnsisting f 100 bservatins in each f tw grups, indicated in blue and in range The purple dashed line represents the Bayes decisin bundary The range backgrund grid indicates the regin in which a test bservatin will be assigned t the range class, and the blue backgrund grid indicates the regin in which a test bservatin will be assigned t the blue class Chapter 5 36 / 52

37 52 Crss validatin Degree=1 Degree=2 Degree=3 Degree=4 Chapter 5 37 / 52

38 52 Crss validatin FIGURE 57 Lgistic regressin fits n the tw-dimensinal classificatin data displayed in Figure 213 The Bayes decisin bundary is represented using a purple dashed line Estimated decisin bundaries frm linear, quadratic, cubic and quartic (degrees 1 t 4) lgistic regressins are displayed in black The (TRUE) test errr rates fr the fur lgistic regressin fits are respectively 0201, 0197, 0160, and 0162, while the Bayes errr rate is 0133 Chapter 5 38 / 52

39 52 Crss validatin Remark abut the simulated example The previus example is simulated The true ppulatin distributin is knwn The figures 0201, 0197, 0160, and 0162, and 0133 (fr Bayes errr rate) are the true test errr, cmputed based n true ppulatin distributin In practice the true ppulatin distributin is unknwn Thus the true test errr cannt be cmputed We use crss validatin t slve the prblem Chapter 5 39 / 52

40 52 Crss validatin Errr Rate Errr Rate Order f Plynmials Used /K Figure: 58 Test errr (brwn), training errr (blue), and 10-fld CV errr (black) n the tw-dimensinal classificatin data displayed in Figure 57 Left: Lgistic regressin using plynmial functins f the predictrs The rder f the plynmials used is displayed n the x-axis Right: The KNN classifier with different values f K, the number f neighbrs used in the KNN classifier Chapter 5 40 / 52

41 52 Crss validatin Training errr declines in general when mdel cmplexity increases Sme times even reaches 0 Test errr general declines first and then increases 10-fld crss validatin prvides reasnable estimate f the test errr, with slight under-estimatin Chapter 5 41 / 52

42 53 Btstrap Bstrap as a resampling prcedure Suppse we have data x 1,, x n, representing the ages f n randmly selected peple in HK Use sample mean x t estimate the ppulatin mean µ, the avearge age f all residents f HK Hw t justify the estimatin errr x µ? Usually by t-cnfidence interval, test f hypthesis They rely n nrmality assumptin r central limit therm Is there anther reliable way? Just btstrap: Chapter 5 42 / 52

43 53 Btstrap Bstrap as a resampling prcedure Take n randm sample (with replacement) frm x 1,, x n calculate the sample mean f the re-sample, dented as x 1 Repeat the abve a large number M times We have x 1, x 2,, x M Use the distributin f x 1 x,, x M x t apprximate that f x µ Chapter 5 43 / 52

44 53 Btstrap Essential idea: Treat the data distribtin (mre prfessinally called empirical distributin) as a prxy f the ppulatin distributin Mimic the data generatin frm the true ppulatin, by trying resampling frm the empirical distributin Mimic yur statistical prcedure (such as cmputing an estimate x) n data, by ding the same n the resampled data Evalute yur statistical prcedure (which may be difficult because it invlves randmness and the unknwn ppulatin distributin) by evaluting yur analgue prcudres n the re-samples Chapter 5 44 / 52

45 53 Btstrap Example X and Y are tw randm variables Then minimizer f var(αx + (1 α)y )) is α = σ 2 Y σ XY σ 2 X + σ2 Y 2σ XY Data: (X 1, Y 1 ),, (X n, Y n ) We can cmpute sample variances and cvariances Estimate α by ˆα = ˆσ 2 Y ˆσ XY ˆσ 2 X + ˆσ2 Y 2ˆσ XY Hw t evalute ˆα α, (remember ˆα is randm and α is unknwn) Use Btstrap Chapter 5 45 / 52

46 53 Btstrap Example Sample n resamples frm (X 1, Y 1 ),, (X n, Y n ), and cmpute the sample the sample variance and cvariances fr this resample And then cmpute ˆα (ˆσ Y = )2 ˆσ XY (ˆσ X )2 + (ˆσ Y )2 2ˆσ XY Repeat this prcedure, and we have ˆα 1,, ˆα M fr a large M Use the distributin f ˆα 1 ˆα,, ˆα M ˆα t apprximate the distributin f ˆα α Fr example, we can use 1 M M (ˆα j ˆα) 2 j=1 t estimate E(ˆα α) 2 Use Btstrap Chapter 5 46 / 52

47 53 Btstrap Y Y X X Y Y X X Figure: 59 Each panel displays 100 simulated returns fr investments X and Y Frm left t right and tp t bttm, the resulting estimates fr α are 0576, 0532, 0657, and 0651 Chapter 5 47 / 52

48 53 Btstrap α α α True Btstrap Figure: 510 Left: A histgram f the estimates f α btained by generating 1,000 simulated data sets frm the true ppulatin Center: A histgram f the estimates f α btained frm 1,000 btstrap samples frm a single data set Right: The estimates f α displayed in the left and center panels are shwn as bxplts In each panel, the pink line indicates the true value f α Chapter 5 48 / 52

49 53 Btstrap Y X Obs Y X Obs Y X Obs Y X Obs Original Data (Z) *1 Z *2 Z Z *B αˆ*1 ˆα *2 ˆα *B Chapter 5 49 / 52

50 53 Btstrap Figure 511 A graphical illustratin f the btstrap apprach n a small sample cntaining n = 3 bservatins Each btstrap data set cntains n bservatins, sampled with replacement frm the riginal data set Each btstrap data set is used t btain an estimate f α Chapter 5 50 / 52

51 53 Btstrap Exercises Exercise 54 f ISLR: Exercises 1-3 and 8 Chapter 5 51 / 52

52 53 Btstrap End f Chapter 5 Chapter 5 52 / 52

Resampling Methods. Cross-validation, Bootstrapping. Marek Petrik 2/21/2017

Resampling Methods. Cross-validation, Bootstrapping. Marek Petrik 2/21/2017 Resampling Methds Crss-validatin, Btstrapping Marek Petrik 2/21/2017 Sme f the figures in this presentatin are taken frm An Intrductin t Statistical Learning, with applicatins in R (Springer, 2013) with

More information

What is Statistical Learning?

What is Statistical Learning? What is Statistical Learning? Sales 5 10 15 20 25 Sales 5 10 15 20 25 Sales 5 10 15 20 25 0 50 100 200 300 TV 0 10 20 30 40 50 Radi 0 20 40 60 80 100 Newspaper Shwn are Sales vs TV, Radi and Newspaper,

More information

Bootstrap Method > # Purpose: understand how bootstrap method works > obs=c(11.96, 5.03, 67.40, 16.07, 31.50, 7.73, 11.10, 22.38) > n=length(obs) >

Bootstrap Method > # Purpose: understand how bootstrap method works > obs=c(11.96, 5.03, 67.40, 16.07, 31.50, 7.73, 11.10, 22.38) > n=length(obs) > Btstrap Methd > # Purpse: understand hw btstrap methd wrks > bs=c(11.96, 5.03, 67.40, 16.07, 31.50, 7.73, 11.10, 22.38) > n=length(bs) > mean(bs) [1] 21.64625 > # estimate f lambda > lambda = 1/mean(bs);

More information

Midwest Big Data Summer School: Machine Learning I: Introduction. Kris De Brabanter

Midwest Big Data Summer School: Machine Learning I: Introduction. Kris De Brabanter Midwest Big Data Summer Schl: Machine Learning I: Intrductin Kris De Brabanter kbrabant@iastate.edu Iwa State University Department f Statistics Department f Cmputer Science June 24, 2016 1/24 Outline

More information

Lecture 2: Supervised vs. unsupervised learning, bias-variance tradeoff

Lecture 2: Supervised vs. unsupervised learning, bias-variance tradeoff Lecture 2: Supervised vs. unsupervised learning, bias-variance tradeff Reading: Chapter 2 STATS 202: Data mining and analysis September 27, 2017 1 / 20 Supervised vs. unsupervised learning In unsupervised

More information

Lecture 2: Supervised vs. unsupervised learning, bias-variance tradeoff

Lecture 2: Supervised vs. unsupervised learning, bias-variance tradeoff Lecture 2: Supervised vs. unsupervised learning, bias-variance tradeff Reading: Chapter 2 STATS 202: Data mining and analysis September 27, 2017 1 / 20 Supervised vs. unsupervised learning In unsupervised

More information

k-nearest Neighbor How to choose k Average of k points more reliable when: Large k: noise in attributes +o o noise in class labels

k-nearest Neighbor How to choose k Average of k points more reliable when: Large k: noise in attributes +o o noise in class labels Mtivating Example Memry-Based Learning Instance-Based Learning K-earest eighbr Inductive Assumptin Similar inputs map t similar utputs If nt true => learning is impssible If true => learning reduces t

More information

Internal vs. external validity. External validity. This section is based on Stock and Watson s Chapter 9.

Internal vs. external validity. External validity. This section is based on Stock and Watson s Chapter 9. Sectin 7 Mdel Assessment This sectin is based n Stck and Watsn s Chapter 9. Internal vs. external validity Internal validity refers t whether the analysis is valid fr the ppulatin and sample being studied.

More information

Pattern Recognition 2014 Support Vector Machines

Pattern Recognition 2014 Support Vector Machines Pattern Recgnitin 2014 Supprt Vectr Machines Ad Feelders Universiteit Utrecht Ad Feelders ( Universiteit Utrecht ) Pattern Recgnitin 1 / 55 Overview 1 Separable Case 2 Kernel Functins 3 Allwing Errrs (Sft

More information

COMP 551 Applied Machine Learning Lecture 4: Linear classification

COMP 551 Applied Machine Learning Lecture 4: Linear classification COMP 551 Applied Machine Learning Lecture 4: Linear classificatin Instructr: Jelle Pineau (jpineau@cs.mcgill.ca) Class web page: www.cs.mcgill.ca/~jpineau/cmp551 Unless therwise nted, all material psted

More information

Comparing Several Means: ANOVA. Group Means and Grand Mean

Comparing Several Means: ANOVA. Group Means and Grand Mean STAT 511 ANOVA and Regressin 1 Cmparing Several Means: ANOVA Slide 1 Blue Lake snap beans were grwn in 12 pen-tp chambers which are subject t 4 treatments 3 each with O 3 and SO 2 present/absent. The ttal

More information

x 1 Outline IAML: Logistic Regression Decision Boundaries Example Data

x 1 Outline IAML: Logistic Regression Decision Boundaries Example Data Outline IAML: Lgistic Regressin Charles Suttn and Victr Lavrenk Schl f Infrmatics Semester Lgistic functin Lgistic regressin Learning lgistic regressin Optimizatin The pwer f nn-linear basis functins Least-squares

More information

Differentiation Applications 1: Related Rates

Differentiation Applications 1: Related Rates Differentiatin Applicatins 1: Related Rates 151 Differentiatin Applicatins 1: Related Rates Mdel 1: Sliding Ladder 10 ladder y 10 ladder 10 ladder A 10 ft ladder is leaning against a wall when the bttm

More information

Simple Linear Regression (single variable)

Simple Linear Regression (single variable) Simple Linear Regressin (single variable) Intrductin t Machine Learning Marek Petrik January 31, 2017 Sme f the figures in this presentatin are taken frm An Intrductin t Statistical Learning, with applicatins

More information

Chapter 3: Cluster Analysis

Chapter 3: Cluster Analysis Chapter 3: Cluster Analysis } 3.1 Basic Cncepts f Clustering 3.1.1 Cluster Analysis 3.1. Clustering Categries } 3. Partitining Methds 3..1 The principle 3.. K-Means Methd 3..3 K-Medids Methd 3..4 CLARA

More information

3.4 Shrinkage Methods Prostate Cancer Data Example (Continued) Ridge Regression

3.4 Shrinkage Methods Prostate Cancer Data Example (Continued) Ridge Regression 3.3.4 Prstate Cancer Data Example (Cntinued) 3.4 Shrinkage Methds 61 Table 3.3 shws the cefficients frm a number f different selectin and shrinkage methds. They are best-subset selectin using an all-subsets

More information

CAUSAL INFERENCE. Technical Track Session I. Phillippe Leite. The World Bank

CAUSAL INFERENCE. Technical Track Session I. Phillippe Leite. The World Bank CAUSAL INFERENCE Technical Track Sessin I Phillippe Leite The Wrld Bank These slides were develped by Christel Vermeersch and mdified by Phillippe Leite fr the purpse f this wrkshp Plicy questins are causal

More information

COMP 551 Applied Machine Learning Lecture 5: Generative models for linear classification

COMP 551 Applied Machine Learning Lecture 5: Generative models for linear classification COMP 551 Applied Machine Learning Lecture 5: Generative mdels fr linear classificatin Instructr: Herke van Hf (herke.vanhf@mail.mcgill.ca) Slides mstly by: Jelle Pineau Class web page: www.cs.mcgill.ca/~hvanh2/cmp551

More information

IAML: Support Vector Machines

IAML: Support Vector Machines 1 / 22 IAML: Supprt Vectr Machines Charles Suttn and Victr Lavrenk Schl f Infrmatics Semester 1 2 / 22 Outline Separating hyperplane with maimum margin Nn-separable training data Epanding the input int

More information

CHAPTER 24: INFERENCE IN REGRESSION. Chapter 24: Make inferences about the population from which the sample data came.

CHAPTER 24: INFERENCE IN REGRESSION. Chapter 24: Make inferences about the population from which the sample data came. MATH 1342 Ch. 24 April 25 and 27, 2013 Page 1 f 5 CHAPTER 24: INFERENCE IN REGRESSION Chapters 4 and 5: Relatinships between tw quantitative variables. Be able t Make a graph (scatterplt) Summarize the

More information

A Matrix Representation of Panel Data

A Matrix Representation of Panel Data web Extensin 6 Appendix 6.A A Matrix Representatin f Panel Data Panel data mdels cme in tw brad varieties, distinct intercept DGPs and errr cmpnent DGPs. his appendix presents matrix algebra representatins

More information

Hypothesis Tests for One Population Mean

Hypothesis Tests for One Population Mean Hypthesis Tests fr One Ppulatin Mean Chapter 9 Ala Abdelbaki Objective Objective: T estimate the value f ne ppulatin mean Inferential statistics using statistics in rder t estimate parameters We will be

More information

Lab 1 The Scientific Method

Lab 1 The Scientific Method INTRODUCTION The fllwing labratry exercise is designed t give yu, the student, an pprtunity t explre unknwn systems, r universes, and hypthesize pssible rules which may gvern the behavir within them. Scientific

More information

STATS216v Introduction to Statistical Learning Stanford University, Summer Practice Final (Solutions) Duration: 3 hours

STATS216v Introduction to Statistical Learning Stanford University, Summer Practice Final (Solutions) Duration: 3 hours STATS216v Intrductin t Statistical Learning Stanfrd University, Summer 2016 Practice Final (Slutins) Duratin: 3 hurs Instructins: (This is a practice final and will nt be graded.) Remember the university

More information

COMP 551 Applied Machine Learning Lecture 11: Support Vector Machines

COMP 551 Applied Machine Learning Lecture 11: Support Vector Machines COMP 551 Applied Machine Learning Lecture 11: Supprt Vectr Machines Instructr: (jpineau@cs.mcgill.ca) Class web page: www.cs.mcgill.ca/~jpineau/cmp551 Unless therwise nted, all material psted fr this curse

More information

Fall 2013 Physics 172 Recitation 3 Momentum and Springs

Fall 2013 Physics 172 Recitation 3 Momentum and Springs Fall 03 Physics 7 Recitatin 3 Mmentum and Springs Purpse: The purpse f this recitatin is t give yu experience wrking with mmentum and the mmentum update frmula. Readings: Chapter.3-.5 Learning Objectives:.3.

More information

A New Evaluation Measure. J. Joiner and L. Werner. The problems of evaluation and the needed criteria of evaluation

A New Evaluation Measure. J. Joiner and L. Werner. The problems of evaluation and the needed criteria of evaluation III-l III. A New Evaluatin Measure J. Jiner and L. Werner Abstract The prblems f evaluatin and the needed criteria f evaluatin measures in the SMART system f infrmatin retrieval are reviewed and discussed.

More information

Module 3: Gaussian Process Parameter Estimation, Prediction Uncertainty, and Diagnostics

Module 3: Gaussian Process Parameter Estimation, Prediction Uncertainty, and Diagnostics Mdule 3: Gaussian Prcess Parameter Estimatin, Predictin Uncertainty, and Diagnstics Jerme Sacks and William J Welch Natinal Institute f Statistical Sciences and University f British Clumbia Adapted frm

More information

AP Statistics Notes Unit Two: The Normal Distributions

AP Statistics Notes Unit Two: The Normal Distributions AP Statistics Ntes Unit Tw: The Nrmal Distributins Syllabus Objectives: 1.5 The student will summarize distributins f data measuring the psitin using quartiles, percentiles, and standardized scres (z-scres).

More information

Kinetic Model Completeness

Kinetic Model Completeness 5.68J/10.652J Spring 2003 Lecture Ntes Tuesday April 15, 2003 Kinetic Mdel Cmpleteness We say a chemical kinetic mdel is cmplete fr a particular reactin cnditin when it cntains all the species and reactins

More information

4th Indian Institute of Astrophysics - PennState Astrostatistics School July, 2013 Vainu Bappu Observatory, Kavalur. Correlation and Regression

4th Indian Institute of Astrophysics - PennState Astrostatistics School July, 2013 Vainu Bappu Observatory, Kavalur. Correlation and Regression 4th Indian Institute f Astrphysics - PennState Astrstatistics Schl July, 2013 Vainu Bappu Observatry, Kavalur Crrelatin and Regressin Rahul Ry Indian Statistical Institute, Delhi. Crrelatin Cnsider a tw

More information

MATCHING TECHNIQUES. Technical Track Session VI. Emanuela Galasso. The World Bank

MATCHING TECHNIQUES. Technical Track Session VI. Emanuela Galasso. The World Bank MATCHING TECHNIQUES Technical Track Sessin VI Emanuela Galass The Wrld Bank These slides were develped by Christel Vermeersch and mdified by Emanuela Galass fr the purpse f this wrkshp When can we use

More information

Statistical Learning. 2.1 What Is Statistical Learning?

Statistical Learning. 2.1 What Is Statistical Learning? 2 Statistical Learning 2.1 What Is Statistical Learning? In rder t mtivate ur study f statistical learning, we begin with a simple example. Suppse that we are statistical cnsultants hired by a client t

More information

Tree Structured Classifier

Tree Structured Classifier Tree Structured Classifier Reference: Classificatin and Regressin Trees by L. Breiman, J. H. Friedman, R. A. Olshen, and C. J. Stne, Chapman & Hall, 98. A Medical Eample (CART): Predict high risk patients

More information

INSTRUMENTAL VARIABLES

INSTRUMENTAL VARIABLES INSTRUMENTAL VARIABLES Technical Track Sessin IV Sergi Urzua University f Maryland Instrumental Variables and IE Tw main uses f IV in impact evaluatin: 1. Crrect fr difference between assignment f treatment

More information

PSU GISPOPSCI June 2011 Ordinary Least Squares & Spatial Linear Regression in GeoDa

PSU GISPOPSCI June 2011 Ordinary Least Squares & Spatial Linear Regression in GeoDa There are tw parts t this lab. The first is intended t demnstrate hw t request and interpret the spatial diagnstics f a standard OLS regressin mdel using GeDa. The diagnstics prvide infrmatin abut the

More information

CHAPTER 4 DIAGNOSTICS FOR INFLUENTIAL OBSERVATIONS

CHAPTER 4 DIAGNOSTICS FOR INFLUENTIAL OBSERVATIONS CHAPTER 4 DIAGNOSTICS FOR INFLUENTIAL OBSERVATIONS 1 Influential bservatins are bservatins whse presence in the data can have a distrting effect n the parameter estimates and pssibly the entire analysis,

More information

Statistics, Numerical Models and Ensembles

Statistics, Numerical Models and Ensembles Statistics, Numerical Mdels and Ensembles Duglas Nychka, Reinhard Furrer,, Dan Cley Claudia Tebaldi, Linda Mearns, Jerry Meehl and Richard Smith (UNC). Spatial predictin and data assimilatin Precipitatin

More information

CS 477/677 Analysis of Algorithms Fall 2007 Dr. George Bebis Course Project Due Date: 11/29/2007

CS 477/677 Analysis of Algorithms Fall 2007 Dr. George Bebis Course Project Due Date: 11/29/2007 CS 477/677 Analysis f Algrithms Fall 2007 Dr. Gerge Bebis Curse Prject Due Date: 11/29/2007 Part1: Cmparisn f Srting Algrithms (70% f the prject grade) The bjective f the first part f the assignment is

More information

Modelling of Clock Behaviour. Don Percival. Applied Physics Laboratory University of Washington Seattle, Washington, USA

Modelling of Clock Behaviour. Don Percival. Applied Physics Laboratory University of Washington Seattle, Washington, USA Mdelling f Clck Behaviur Dn Percival Applied Physics Labratry University f Washingtn Seattle, Washingtn, USA verheads and paper fr talk available at http://faculty.washingtn.edu/dbp/talks.html 1 Overview

More information

5 th grade Common Core Standards

5 th grade Common Core Standards 5 th grade Cmmn Cre Standards In Grade 5, instructinal time shuld fcus n three critical areas: (1) develping fluency with additin and subtractin f fractins, and develping understanding f the multiplicatin

More information

the results to larger systems due to prop'erties of the projection algorithm. First, the number of hidden nodes must

the results to larger systems due to prop'erties of the projection algorithm. First, the number of hidden nodes must M.E. Aggune, M.J. Dambrg, M.A. El-Sharkawi, R.J. Marks II and L.E. Atlas, "Dynamic and static security assessment f pwer systems using artificial neural netwrks", Prceedings f the NSF Wrkshp n Applicatins

More information

AP Statistics Practice Test Unit Three Exploring Relationships Between Variables. Name Period Date

AP Statistics Practice Test Unit Three Exploring Relationships Between Variables. Name Period Date AP Statistics Practice Test Unit Three Explring Relatinships Between Variables Name Perid Date True r False: 1. Crrelatin and regressin require explanatry and respnse variables. 1. 2. Every least squares

More information

Smoothing, penalized least squares and splines

Smoothing, penalized least squares and splines Smthing, penalized least squares and splines Duglas Nychka, www.image.ucar.edu/~nychka Lcally weighted averages Penalized least squares smthers Prperties f smthers Splines and Reprducing Kernels The interplatin

More information

Math Foundations 20 Work Plan

Math Foundations 20 Work Plan Math Fundatins 20 Wrk Plan Units / Tpics 20.8 Demnstrate understanding f systems f linear inequalities in tw variables. Time Frame December 1-3 weeks 6-10 Majr Learning Indicatrs Identify situatins relevant

More information

You need to be able to define the following terms and answer basic questions about them:

You need to be able to define the following terms and answer basic questions about them: CS440/ECE448 Sectin Q Fall 2017 Midterm Review Yu need t be able t define the fllwing terms and answer basic questins abut them: Intr t AI, agents and envirnments Pssible definitins f AI, prs and cns f

More information

SUPPLEMENTARY MATERIAL GaGa: a simple and flexible hierarchical model for microarray data analysis

SUPPLEMENTARY MATERIAL GaGa: a simple and flexible hierarchical model for microarray data analysis SUPPLEMENTARY MATERIAL GaGa: a simple and flexible hierarchical mdel fr micrarray data analysis David Rssell Department f Bistatistics M.D. Andersn Cancer Center, Hustn, TX 77030, USA rsselldavid@gmail.cm

More information

Interference is when two (or more) sets of waves meet and combine to produce a new pattern.

Interference is when two (or more) sets of waves meet and combine to produce a new pattern. Interference Interference is when tw (r mre) sets f waves meet and cmbine t prduce a new pattern. This pattern can vary depending n the riginal wave directin, wavelength, amplitude, etc. The tw mst extreme

More information

, which yields. where z1. and z2

, which yields. where z1. and z2 The Gaussian r Nrmal PDF, Page 1 The Gaussian r Nrmal Prbability Density Functin Authr: Jhn M Cimbala, Penn State University Latest revisin: 11 September 13 The Gaussian r Nrmal Prbability Density Functin

More information

MATCHING TECHNIQUES Technical Track Session VI Céline Ferré The World Bank

MATCHING TECHNIQUES Technical Track Session VI Céline Ferré The World Bank MATCHING TECHNIQUES Technical Track Sessin VI Céline Ferré The Wrld Bank When can we use matching? What if the assignment t the treatment is nt dne randmly r based n an eligibility index, but n the basis

More information

2004 AP CHEMISTRY FREE-RESPONSE QUESTIONS

2004 AP CHEMISTRY FREE-RESPONSE QUESTIONS 2004 AP CHEMISTRY FREE-RESPONSE QUESTIONS 6. An electrchemical cell is cnstructed with an pen switch, as shwn in the diagram abve. A strip f Sn and a strip f an unknwn metal, X, are used as electrdes.

More information

ENSC Discrete Time Systems. Project Outline. Semester

ENSC Discrete Time Systems. Project Outline. Semester ENSC 49 - iscrete Time Systems Prject Outline Semester 006-1. Objectives The gal f the prject is t design a channel fading simulatr. Upn successful cmpletin f the prject, yu will reinfrce yur understanding

More information

Localized Model Selection for Regression

Localized Model Selection for Regression Lcalized Mdel Selectin fr Regressin Yuhng Yang Schl f Statistics University f Minnesta Church Street S.E. Minneaplis, MN 5555 May 7, 007 Abstract Research n mdel/prcedure selectin has fcused n selecting

More information

The general linear model and Statistical Parametric Mapping I: Introduction to the GLM

The general linear model and Statistical Parametric Mapping I: Introduction to the GLM The general linear mdel and Statistical Parametric Mapping I: Intrductin t the GLM Alexa Mrcm and Stefan Kiebel, Rik Hensn, Andrew Hlmes & J-B J Pline Overview Intrductin Essential cncepts Mdelling Design

More information

Chapter 15 & 16: Random Forests & Ensemble Learning

Chapter 15 & 16: Random Forests & Ensemble Learning Chapter 15 & 16: Randm Frests & Ensemble Learning DD3364 Nvember 27, 2012 Ty Prblem fr Bsted Tree Bsted Tree Example Estimate this functin with a sum f trees with 9-terminal ndes by minimizing the sum

More information

SIZE BIAS IN LINE TRANSECT SAMPLING: A FIELD TEST. Mark C. Otto Statistics Research Division, Bureau of the Census Washington, D.C , U.S.A.

SIZE BIAS IN LINE TRANSECT SAMPLING: A FIELD TEST. Mark C. Otto Statistics Research Division, Bureau of the Census Washington, D.C , U.S.A. SIZE BIAS IN LINE TRANSECT SAMPLING: A FIELD TEST Mark C. Ott Statistics Research Divisin, Bureau f the Census Washingtn, D.C. 20233, U.S.A. and Kenneth H. Pllck Department f Statistics, Nrth Carlina State

More information

Five Whys How To Do It Better

Five Whys How To Do It Better Five Whys Definitin. As explained in the previus article, we define rt cause as simply the uncvering f hw the current prblem came int being. Fr a simple causal chain, it is the entire chain. Fr a cmplex

More information

Distributions, spatial statistics and a Bayesian perspective

Distributions, spatial statistics and a Bayesian perspective Distributins, spatial statistics and a Bayesian perspective Dug Nychka Natinal Center fr Atmspheric Research Distributins and densities Cnditinal distributins and Bayes Thm Bivariate nrmal Spatial statistics

More information

Activity Guide Loops and Random Numbers

Activity Guide Loops and Random Numbers Unit 3 Lessn 7 Name(s) Perid Date Activity Guide Lps and Randm Numbers CS Cntent Lps are a relatively straightfrward idea in prgramming - yu want a certain chunk f cde t run repeatedly - but it takes a

More information

SAMPLING DYNAMICAL SYSTEMS

SAMPLING DYNAMICAL SYSTEMS SAMPLING DYNAMICAL SYSTEMS Melvin J. Hinich Applied Research Labratries The University f Texas at Austin Austin, TX 78713-8029, USA (512) 835-3278 (Vice) 835-3259 (Fax) hinich@mail.la.utexas.edu ABSTRACT

More information

Biplots in Practice MICHAEL GREENACRE. Professor of Statistics at the Pompeu Fabra University. Chapter 13 Offprint

Biplots in Practice MICHAEL GREENACRE. Professor of Statistics at the Pompeu Fabra University. Chapter 13 Offprint Biplts in Practice MICHAEL GREENACRE Prfessr f Statistics at the Pmpeu Fabra University Chapter 13 Offprint CASE STUDY BIOMEDICINE Cmparing Cancer Types Accrding t Gene Epressin Arrays First published:

More information

A Regression Solution to the Problem of Criterion Score Comparability

A Regression Solution to the Problem of Criterion Score Comparability A Regressin Slutin t the Prblem f Criterin Scre Cmparability William M. Pugh Naval Health Research Center When the criterin measure in a study is the accumulatin f respnses r behavirs fr an individual

More information

Inference in the Multiple-Regression

Inference in the Multiple-Regression Sectin 5 Mdel Inference in the Multiple-Regressin Kinds f hypthesis tests in a multiple regressin There are several distinct kinds f hypthesis tests we can run in a multiple regressin. Suppse that amng

More information

CN700 Additive Models and Trees Chapter 9: Hastie et al. (2001)

CN700 Additive Models and Trees Chapter 9: Hastie et al. (2001) CN700 Additive Mdels and Trees Chapter 9: Hastie et al. (2001) Madhusudana Shashanka Department f Cgnitive and Neural Systems Bstn University CN700 - Additive Mdels and Trees March 02, 2004 p.1/34 Overview

More information

NAME: Prof. Ruiz. 1. [5 points] What is the difference between simple random sampling and stratified random sampling?

NAME: Prof. Ruiz. 1. [5 points] What is the difference between simple random sampling and stratified random sampling? CS4445 ata Mining and Kwledge iscery in atabases. B Term 2014 Exam 1 Nember 24, 2014 Prf. Carlina Ruiz epartment f Cmputer Science Wrcester Plytechnic Institute NAME: Prf. Ruiz Prblem I: Prblem II: Prblem

More information

Support Vector Machines and Flexible Discriminants

Support Vector Machines and Flexible Discriminants 12 Supprt Vectr Machines and Flexible Discriminants This is page 417 Printer: Opaque this 12.1 Intrductin In this chapter we describe generalizatins f linear decisin bundaries fr classificatin. Optimal

More information

T Algorithmic methods for data mining. Slide set 6: dimensionality reduction

T Algorithmic methods for data mining. Slide set 6: dimensionality reduction T-61.5060 Algrithmic methds fr data mining Slide set 6: dimensinality reductin reading assignment LRU bk: 11.1 11.3 PCA tutrial in mycurses (ptinal) ptinal: An Elementary Prf f a Therem f Jhnsn and Lindenstrauss,

More information

Part 3 Introduction to statistical classification techniques

Part 3 Introduction to statistical classification techniques Part 3 Intrductin t statistical classificatin techniques Machine Learning, Part 3, March 07 Fabi Rli Preamble ØIn Part we have seen that if we knw: Psterir prbabilities P(ω i / ) Or the equivalent terms

More information

On Out-of-Sample Statistics for Financial Time-Series

On Out-of-Sample Statistics for Financial Time-Series On Out-f-Sample Statistics fr Financial Time-Series Françis Gingras Yshua Bengi Claude Nadeau CRM-2585 January 1999 Département de physique, Université de Mntréal Labratire d infrmatique des systèmes adaptatifs,

More information

NUMBERS, MATHEMATICS AND EQUATIONS

NUMBERS, MATHEMATICS AND EQUATIONS AUSTRALIAN CURRICULUM PHYSICS GETTING STARTED WITH PHYSICS NUMBERS, MATHEMATICS AND EQUATIONS An integral part t the understanding f ur physical wrld is the use f mathematical mdels which can be used t

More information

Linear programming III

Linear programming III Linear prgramming III Review 1/33 What have cvered in previus tw classes LP prblem setup: linear bjective functin, linear cnstraints. exist extreme pint ptimal slutin. Simplex methd: g thrugh extreme pint

More information

We say that y is a linear function of x if. Chapter 13: The Correlation Coefficient and the Regression Line

We say that y is a linear function of x if. Chapter 13: The Correlation Coefficient and the Regression Line Chapter 13: The Crrelatin Cefficient and the Regressin Line We begin with a sme useful facts abut straight lines. Recall the x, y crdinate system, as pictured belw. 3 2 1 y = 2.5 y = 0.5x 3 2 1 1 2 3 1

More information

and the Doppler frequency rate f R , can be related to the coefficients of this polynomial. The relationships are:

and the Doppler frequency rate f R , can be related to the coefficients of this polynomial. The relationships are: Algrithm fr Estimating R and R - (David Sandwell, SIO, August 4, 2006) Azimith cmpressin invlves the alignment f successive eches t be fcused n a pint target Let s be the slw time alng the satellite track

More information

Lecture 10, Principal Component Analysis

Lecture 10, Principal Component Analysis Principal Cmpnent Analysis Lecture 10, Principal Cmpnent Analysis Ha Helen Zhang Fall 2017 Ha Helen Zhang Lecture 10, Principal Cmpnent Analysis 1 / 16 Principal Cmpnent Analysis Lecture 10, Principal

More information

Name: Block: Date: Science 10: The Great Geyser Experiment A controlled experiment

Name: Block: Date: Science 10: The Great Geyser Experiment A controlled experiment Science 10: The Great Geyser Experiment A cntrlled experiment Yu will prduce a GEYSER by drpping Ments int a bttle f diet pp Sme questins t think abut are: What are yu ging t test? What are yu ging t measure?

More information

Experiment #3. Graphing with Excel

Experiment #3. Graphing with Excel Experiment #3. Graphing with Excel Study the "Graphing with Excel" instructins that have been prvided. Additinal help with learning t use Excel can be fund n several web sites, including http://www.ncsu.edu/labwrite/res/gt/gt-

More information

I.S. 239 Mark Twain. Grade 7 Mathematics Spring Performance Task: Proportional Relationships

I.S. 239 Mark Twain. Grade 7 Mathematics Spring Performance Task: Proportional Relationships I.S. 239 Mark Twain 7 ID Name: Date: Grade 7 Mathematics Spring Perfrmance Task: Prprtinal Relatinships Directins: Cmplete all parts f each sheet fr each given task. Be sure t read thrugh the rubrics s

More information

Determining the Accuracy of Modal Parameter Estimation Methods

Determining the Accuracy of Modal Parameter Estimation Methods Determining the Accuracy f Mdal Parameter Estimatin Methds by Michael Lee Ph.D., P.E. & Mar Richardsn Ph.D. Structural Measurement Systems Milpitas, CA Abstract The mst cmmn type f mdal testing system

More information

https://goo.gl/eaqvfo SUMMER REV: Half-Life DUE DATE: JULY 2 nd

https://goo.gl/eaqvfo SUMMER REV: Half-Life DUE DATE: JULY 2 nd NAME: DUE DATE: JULY 2 nd AP Chemistry SUMMER REV: Half-Life Why? Every radiistpe has a characteristic rate f decay measured by its half-life. Half-lives can be as shrt as a fractin f a secnd r as lng

More information

Discussion on Regularized Regression for Categorical Data (Tutz and Gertheiss)

Discussion on Regularized Regression for Categorical Data (Tutz and Gertheiss) Discussin n Regularized Regressin fr Categrical Data (Tutz and Gertheiss) Peter Bühlmann, Ruben Dezeure Seminar fr Statistics, Department f Mathematics, ETH Zürich, Switzerland Address fr crrespndence:

More information

Perfrmance f Sensitizing Rules n Shewhart Cntrl Charts with Autcrrelated Data Key Wrds: Autregressive, Mving Average, Runs Tests, Shewhart Cntrl Chart

Perfrmance f Sensitizing Rules n Shewhart Cntrl Charts with Autcrrelated Data Key Wrds: Autregressive, Mving Average, Runs Tests, Shewhart Cntrl Chart Perfrmance f Sensitizing Rules n Shewhart Cntrl Charts with Autcrrelated Data Sandy D. Balkin Dennis K. J. Lin y Pennsylvania State University, University Park, PA 16802 Sandy Balkin is a graduate student

More information

In SMV I. IAML: Support Vector Machines II. This Time. The SVM optimization problem. We saw:

In SMV I. IAML: Support Vector Machines II. This Time. The SVM optimization problem. We saw: In SMV I IAML: Supprt Vectr Machines II Nigel Gddard Schl f Infrmatics Semester 1 We sa: Ma margin trick Gemetry f the margin and h t cmpute it Finding the ma margin hyperplane using a cnstrained ptimizatin

More information

CHAPTER 3 INEQUALITIES. Copyright -The Institute of Chartered Accountants of India

CHAPTER 3 INEQUALITIES. Copyright -The Institute of Chartered Accountants of India CHAPTER 3 INEQUALITIES Cpyright -The Institute f Chartered Accuntants f India INEQUALITIES LEARNING OBJECTIVES One f the widely used decisin making prblems, nwadays, is t decide n the ptimal mix f scarce

More information

Introduction to Regression

Introduction to Regression Intrductin t Regressin Administrivia Hmewrk 6 psted later tnight. Due Friday after Break. 2 Statistical Mdeling Thus far we ve talked abut Descriptive Statistics: This is the way my sample is Inferential

More information

This section is primarily focused on tools to aid us in finding roots/zeros/ -intercepts of polynomials. Essentially, our focus turns to solving.

This section is primarily focused on tools to aid us in finding roots/zeros/ -intercepts of polynomials. Essentially, our focus turns to solving. Sectin 3.2: Many f yu WILL need t watch the crrespnding vides fr this sectin n MyOpenMath! This sectin is primarily fcused n tls t aid us in finding rts/zers/ -intercepts f plynmials. Essentially, ur fcus

More information

Churn Prediction using Dynamic RFM-Augmented node2vec

Churn Prediction using Dynamic RFM-Augmented node2vec Churn Predictin using Dynamic RFM-Augmented nde2vec Sandra Mitrvić, Jchen de Weerdt, Bart Baesens & Wilfried Lemahieu Department f Decisin Sciences and Infrmatin Management, KU Leuven 18 September 2017,

More information

Stats Classification Ji Zhu, Michigan Statistics 1. Classification. Ji Zhu 445C West Hall

Stats Classification Ji Zhu, Michigan Statistics 1. Classification. Ji Zhu 445C West Hall Stats 415 - Classificatin Ji Zhu, Michigan Statistics 1 Classificatin Ji Zhu 445C West Hall 734-936-2577 jizhu@umich.edu Stats 415 - Classificatin Ji Zhu, Michigan Statistics 2 Examples f Classificatin

More information

CHM112 Lab Graphing with Excel Grading Rubric

CHM112 Lab Graphing with Excel Grading Rubric Name CHM112 Lab Graphing with Excel Grading Rubric Criteria Pints pssible Pints earned Graphs crrectly pltted and adhere t all guidelines (including descriptive title, prperly frmatted axes, trendline

More information

ON-LINE PROCEDURE FOR TERMINATING AN ACCELERATED DEGRADATION TEST

ON-LINE PROCEDURE FOR TERMINATING AN ACCELERATED DEGRADATION TEST Statistica Sinica 8(1998), 207-220 ON-LINE PROCEDURE FOR TERMINATING AN ACCELERATED DEGRADATION TEST Hng-Fwu Yu and Sheng-Tsaing Tseng Natinal Taiwan University f Science and Technlgy and Natinal Tsing-Hua

More information

Lead/Lag Compensator Frequency Domain Properties and Design Methods

Lead/Lag Compensator Frequency Domain Properties and Design Methods Lectures 6 and 7 Lead/Lag Cmpensatr Frequency Dmain Prperties and Design Methds Definitin Cnsider the cmpensatr (ie cntrller Fr, it is called a lag cmpensatr s K Fr s, it is called a lead cmpensatr Ntatin

More information

MODULE FOUR. This module addresses functions. SC Academic Elementary Algebra Standards:

MODULE FOUR. This module addresses functions. SC Academic Elementary Algebra Standards: MODULE FOUR This mdule addresses functins SC Academic Standards: EA-3.1 Classify a relatinship as being either a functin r nt a functin when given data as a table, set f rdered pairs, r graph. EA-3.2 Use

More information

Weathering. Title: Chemical and Mechanical Weathering. Grade Level: Subject/Content: Earth and Space Science

Weathering. Title: Chemical and Mechanical Weathering. Grade Level: Subject/Content: Earth and Space Science Weathering Title: Chemical and Mechanical Weathering Grade Level: 9-12 Subject/Cntent: Earth and Space Science Summary f Lessn: Students will test hw chemical and mechanical weathering can affect a rck

More information

Computational modeling techniques

Computational modeling techniques Cmputatinal mdeling techniques Lecture 2: Mdeling change. In Petre Department f IT, Åb Akademi http://users.ab.fi/ipetre/cmpmd/ Cntent f the lecture Basic paradigm f mdeling change Examples Linear dynamical

More information

Medium Scale Integrated (MSI) devices [Sections 2.9 and 2.10]

Medium Scale Integrated (MSI) devices [Sections 2.9 and 2.10] EECS 270, Winter 2017, Lecture 3 Page 1 f 6 Medium Scale Integrated (MSI) devices [Sectins 2.9 and 2.10] As we ve seen, it s smetimes nt reasnable t d all the design wrk at the gate-level smetimes we just

More information

LHS Mathematics Department Honors Pre-Calculus Final Exam 2002 Answers

LHS Mathematics Department Honors Pre-Calculus Final Exam 2002 Answers LHS Mathematics Department Hnrs Pre-alculus Final Eam nswers Part Shrt Prblems The table at the right gives the ppulatin f Massachusetts ver the past several decades Using an epnential mdel, predict the

More information

Eric Klein and Ning Sa

Eric Klein and Ning Sa Week 12. Statistical Appraches t Netwrks: p1 and p* Wasserman and Faust Chapter 15: Statistical Analysis f Single Relatinal Netwrks There are fur tasks in psitinal analysis: 1) Define Equivalence 2) Measure

More information

IN a recent article, Geary [1972] discussed the merit of taking first differences

IN a recent article, Geary [1972] discussed the merit of taking first differences The Efficiency f Taking First Differences in Regressin Analysis: A Nte J. A. TILLMAN IN a recent article, Geary [1972] discussed the merit f taking first differences t deal with the prblems that trends

More information

Chapter Summary. Mathematical Induction Strong Induction Recursive Definitions Structural Induction Recursive Algorithms

Chapter Summary. Mathematical Induction Strong Induction Recursive Definitions Structural Induction Recursive Algorithms Chapter 5 1 Chapter Summary Mathematical Inductin Strng Inductin Recursive Definitins Structural Inductin Recursive Algrithms Sectin 5.1 3 Sectin Summary Mathematical Inductin Examples f Prf by Mathematical

More information

Acids and Bases Lesson 3

Acids and Bases Lesson 3 Acids and Bases Lessn 3 The ph f a slutin is defined as the negative lgarithm, t the base ten, f the hydrnium in cncentratin. In a neutral slutin at 25 C, the hydrnium in and the hydrxide in cncentratins

More information

Reinforcement Learning" CMPSCI 383 Nov 29, 2011!

Reinforcement Learning CMPSCI 383 Nov 29, 2011! Reinfrcement Learning" CMPSCI 383 Nv 29, 2011! 1 Tdayʼs lecture" Review f Chapter 17: Making Cmple Decisins! Sequential decisin prblems! The mtivatin and advantages f reinfrcement learning.! Passive learning!

More information