Localized Model Selection for Regression

Size: px
Start display at page:

Download "Localized Model Selection for Regression"

Transcription

1 Lcalized Mdel Selectin fr Regressin Yuhng Yang Schl f Statistics University f Minnesta Church Street S.E. Minneaplis, MN 5555 May 7, 007 Abstract Research n mdel/prcedure selectin has fcused n selecting a single mdel glbally. In many applicatins, especially fr high-dimensinal r cmplex data, hwever, the relative perfrmance f the candidate prcedures typically depends n the lcatin, and the glbally best prcedure can ften be imprved when selectin f a mdel is allwed t depend n lcatin. We cnsider lcalized mdel selectin and derive their theretical prperties. 1 Intrductin In statistical mdeling, usually a number f candidate mdels are cnsidered. Traditinal mdel selectin thery and practice fcus n the search f a single candidate that is treated as the true r best mdel. There are gd reasns fr this: 1 cmparing the candidates in terms f glbal perfrmance is a gd starting pint; in simple situatins, the relative perfrmance in ranking f the candidate mdels may nt depend n x the vectr f predictrs; 3 if ne f the candidate mdels is believed t describe the data very well, ne des want t find it fr ptimal predictin and interpretatin. Hwever, fr high dimensinal r cmplex data, glbal selectin may be sub-ptimal. Cnsider tw example situatins. 1. Fr univariate regressin, if the mean functin is infinite dimensinal with respect t the candidate mdels, there des nt seem t be a strng reasn t believe that ne mdel wrks better than the thers at all x values. Thus a glbal ranking and selectin may nt be the best thing t d. This will be demnstrated later.. In many current statistical applicatins, the number f predictrs is very large, r even much larger than the number f bservatins. In such cases, there can be substantial uncertainty in mdeling. Fr instance, when selecting imprtant genes fr explaining a respnse variable ut f 5000 genes based n 100 bservatins, any variable selectin methd is explratry in nature and it seems 1

2 clear that ne cannt expect the selected mdel t have really captured the relatinship between the respnse and the predictrs. When ne cnsiders different types f mdels, they ften perfrm the best in different regins. The abve situatins mtivate the cnsideratin f lcalized mdel selectin. With the ever increasing cmputing pwer, selecting prcedures lcally becmes feasible in implementatin. Given the directin f lcalized mdel selectin, ne may wnder if it is better t take a lcal estimatin apprach in the first place and put the effrt n building a single gd estimatr. In ur pinin, this des nt wrk in general. First, glbal cnsideratins are imprtant fr aviding verfit, and fr high dimensinal situatins, lcal estimatin withut glbal cnstraints is ften nt pssible. Secnd, ne has many different ways t d lcal estimatin and then ne is back t the prblem f prcedure selectin. Nw let us set up the prblem mathematically. Let X i, Y i, i = 1,..., n be iid bservatins with X i taking values in X, a measurable set in R d fr sme d 1. Let Y i = fx i ε i, where f is the regressin functin t be estimated under squared errr lss and the errr ε i has mean zer and variance σ. Unless stated therwise, ε i is assumed t be independent f X i. Suppse δ j, j J are a finite cllectin f statistical prcedures fr estimating the regressin functin, each prducing an estimatr f f based n a given sample. We will fcus n the case with tw prcedures, althugh similar results hld mre generally. The rest f the paper is rganized as fllws. In Sectin, we give an illustratin t mtivate lcalized mdel selectin. In Sectin 3, we mentin three appraches t lcalized mdel selectin. In Sectin, we prvide a result that characterizes perfrmance f a lcalized crss validatin methd. In Sectin 5, we study preference-regin-based lcalized mdel selectin. Cncluding remarks are given in Sectin 6. Prfs f the therems are put in an appendix. A mtivating illustratin Cnsider estimating a regressin functin n [ 1, 1] based n 100 bservatins. The x values are unifrmly spaced. The true regressin functin is fx = 0.5x 0.8 π exp 00x π exp 00x 0.5 and the errr is frm N0, 0.3. A typical realizatin f the data is given in Figure 1. The linear trend is bvius, but ne als sees pssible nnlinearities. Naturally, ne may cnsider a simple linear mdel thinking that the smewhat

3 unusual pattern in the middle might be caused by a few utliers r the deviatin frm linearity may nt be serius enugh t pursue. Alternatively ne may cnsider a nnparametric methd. We chse a smthing spline methd that is prvided in R with the default chices f the cntrl parameters y A Typical Realizatin f 100 Observatins frm the Nn linear Mdel x Figure 1: Scatter Plt f a Typical Realizatin f 100 Observatins.1 Advantage f lcalized selectin with a certain frm f preference regin We cmpare a crss validatin CV methd fr glbal selectin between the linear regressin and the smthing spline prcedures with a selectin methd that recgnizes that the tw estimatrs may perfrm well in different regins. Fr the CV methd, we randmly split the data int tw equally sized parts, find the linear and the smthing spline SS estimates using the first part and cmpute the predictin squared errr n the secnd part. We repeat this 50 times and chse the ne with smaller median predictin errr ver the randm splittings. Let f G x be the resulting estimatr. Frm the scatter plt, ne may suspect that the linear estimate may perfrm prly arund 0, while the nnparametric estimate may be undersmth when x is away frm 0. Thus, fr the cmpeting nn-glbal selectin methd, we cnsider estimatrs f the frm fx; c = f L xi { x c} f SS xi { x <c}, where f L is the estimatr f f based n the linear mdel, and f SS is the SS estimatr. We use CV similarly as befre t chse c in the range f [0, 1] at a grid f width Let f NG x be the resulting estimatr. Nte that when c = 0, the linear estimatr is selected fr all x, and when c = 1, the 3

4 nnparametric estimatr is selected fr all x, but fr c between, we use SS nly when x is n bigger than c. At a given x 0, cmpute the squared errr lsses f f G x 0 and f NG x 0 respectively. Generate 00 independent data sets frm the true mdel t simulate the risks f f G x 0 and f NG x 0. The results are presented in Figure at a number f x 0 values in the range f 1, 1. Cmparing Risks: Glbal Selectin vs Lcal Selectin Risk Rati x Figure : Risk Rati f a Glbal CV vs a Nn-Glbal Selectin Methd The figure clearly shws that the nn-glbal selectin perfrms much better than the glbal selectin except in a very small neighbrhd arund zer, demnstrating the ptentially great advantage f cnsidering nn-glbal selectin amng candidate prcedures. It is prbably fair t say that ne may nt necessarily chse the frm f fx; c frm inspecting the scatter plt. In the next subsectin, we use lgistic regressin t find the preference regin autmatically.. Classificatin fr estimating the preference regin We cntinue with the earlier setting except that the sample size is nw 00 s that the cntrast is mre clearly seen in the next figure. We fcus n ne typical realizatin f the data. A scatter plt f the data with the linear and the smthing spline fits is given in the upper-left panel f Figure 3. We use lgistic regressin as the classificatin methd t find the regin where the linear estimatr perfrms better than the smthing spline estimatr. With a randm splitting f the data int tw parts f equal size, we fit a straight-line mdel and the smthing spline using the first 100 bservatins, and btain the binary variable that indicates which methd has a smaller predictin errr n the secnd 100 bservatins. A

5 typical utcme is in the upper-right panel f Figure 3. One may get the impressin that SS tends t d better in the middle and less well at the ends. We fit a lgistic regressin mdel with three terms: 1, x and x. The estimated prbability that the linear mdel perfrms better at x is px = 1 1 exp x 1.97x. Nte that px > 0.5 crrespnds t x < 0.51 r x > 0.9, which is very sensible judging frm ur knwledge f the true mean functin despite that the upper-right panel f the figure des nt seem t be very visually infrmative fr relating the relative perfrmance f the tw estimatrs and x. This yields the fllwing cmbined estimate f the regressin functin { ˆfL x if x < 0.51 r x > 0.9 fx = ˆf SS x therwise, where ˆf L x and ˆf SS x are the linear and smthing spline estimates f f based n the full data. This estimate is cmpared with the true functin in the lwer-right panel f Figure 3. Frm the lwer-left panel f Figure 3, we see that the linear estimate is very pr arund -0.5 and 0.5 nt surprisingly; in cntrast, the SS estimate is quite gd there, but it pays a price fr being flexible in terms f accuracy at ther lcatins. Glbally, the SS is much better than the linear estimate. Indeed, the rati f their integrated squared errrs is ver 6.3. Thus frm the glbal perspective, the SS estimate is the clear winner. Hwever, with the use f lgistic regressin, we prperly figured ut that when x is far away frm zer by a certain degree, the linear estimate is better. Althugh the linear estimate is slightly biased even in the linear parts, its use in the final cmbined estimate is very helpful as seen in the figure. Numerically, the integrated squared errr f the cmbined estimate is nly 35% f that f the SS estimate. Therefre a glbally pr methd can still make a gd cntributin if it is prperly cmbined with a glbally much better estimatr. 3 Three appraches t nn-glbal mdel selectin Cnsider tw prcedures δ 1 and δ with risks Rδ 1 ; x; n and Rδ ; x; n at a given x value based n a sample f size n. Let A = {x : Rδ 1 ; x; n Rδ ; x; n} be the set f x at which the first prcedure perfrms n wrse than the secnd ne. It is the preference regin f δ 1 relative t δ. Ideally, fr selectin, ne wuld use δ 1 n A and δ n A c. In reality, ne may have little prir knwledge n the frm f A and t deal with the issue, ne may cnsider graphical inspectins when x is f a lw dimensin r cnsider A frm a class f sets with a prper cmplexity in hpe that ne member is clse t A. One can cnsider varius sets f A f different degrees f lcality. We cnsider three appraches 5

6 Scatter Plt with Fitted Values Abs. Pred. Err. n Test Data y Abslute Predictin Errr Lin.: SS: x x Cmpare Estimates with True Mean Cmpare Cmbined Estimate with True Mean True Mean and Estimates True Lin. SS True Mean and Cmbined Estimate True Cmb x x Figure 3: An Example f Using Lgistic Regressin fr Nn-Glbal Selectin belw. 3.1 Neighbrhd based selectin Instead f directly estimating A, at each given x 0, ne cnsiders a lcal neighbrhd arund x 0 and tries t find the candidate that perfrms better in the lcal area. This apprach can be very helpful when A cannt be described well by simple sets. It will be studied in detail in the next sectin. 6

7 3. Empirical risk minimizatin One cnsiders a cllectin f sets f a certain mathematical frm and tries t identify the ne with best perfrmance. Here the cllectin may depend n the sample size. The size f the cllectin can be pre-determined r adaptively chsen. This apprach will be briefly explred in Sectin Classificatin based selectin Smetimes, ne is less cnfident t g with any specific cllectin f sets as the preference regin. This is especially true when the input dimensin is high, where neighbrhd based selectin may perfrm prly due t the curse f dimensinality. In such a case, ne can take advantage f classificatin methds t cnduct lcalized selectin as in Sectin.. In general, ne splits the data int tw parts. The first part is used t btain the estimates frm the candidate prcedures, and then make predictins fr the secnd part. Based n the relative predictive perfrmance in each case, we create a new variable that simply indicates which estimate is the better ne. Then we can apply a sensible classificatin methd t relate the perfrmance indicatin variable t the cvariates. If a classifier perfrms well, it indicates that the candidate prcedures relative perfrmance depends n the cvariates, and we can d better than glbally selecting ne f the prcedures. Lcalized crss validatin selectin Crss validatin see, e.g., Allen, 197; Stne, 197; Geisser, 1975 has been widely used fr mdel selectin. In this sectin, we cnsider a lcalized crss validatin methd fr selecting a candidate prcedure lcally. Fr a given x 0, cnsider the ball centered at x 0 with radius r n fr sme r n > 0 under the Euclidean distance. We randmly split the data int a training set f size n 1 and a test set f size n, and then use the training set t d estimatin by each regressin prcedure. Fr evaluatin, cnsider nly the data pints in the test set that are in the given neighbrhd f x 0. Let ĵx 0 = ĵ n x 0 be the prcedure that has the smaller average squared predictin errr. This prcess is repeated with a number f randm splittings f the bservatins t avid the splitting bias. The prcedure that wins mre frequently ver the permutatins is the final winner. We call this a lcalized crss validatin L-CV at x 0 f r n -neighbrhd. Fr the lcal selectin prblem, we are interested in the questin that under what cnditins, the lcally better estimatr will be selected with high prbability. Mre precisely, assuming that when n is large enugh, ne prcedure has risk at x 0 smaller than that f the ther ne, we want t select the 7

8 better ne with prbability appraching 1. If a selectin methd has this prperty, we say it is lcally cnsistent in selectin at x 0. Fr related results n glbal CV, see, e.g., Li 1987, Burman 1989, Zhang 1993, Sha 1997, Wegkamp 003, and Yang 007. Given x 0 and η > 0, let L f; x 0 ; η = x Bx 0;η fx fx P ex dx be a lcal lss f an estimate f arund x 0, where P ex dentes the distributin f X cnditinal n that X takes value in the neighbrhd Bx 0 ; η. Let s,x0;η dente the L s nrm arund x 0 with radius η, i.e., g s,x0;η= x Bx0;η gx s P ex dx 1/s. Definitin 1. Prcedure δ 1 r f 1,n is asympttically better than δ r f,n under the squared lss at η n -neighbrhd f x 0, dented δ 1 δ at x 0 ; η n, if fr each nn-increasing sequence η n with 0 < η n η n and every 0 < ǫ < 1, there exists a cnstant c ǫ > 0 such that when n is large enugh, P L f,n ; x 0 ; η n 1 c ǫ L f 1,n ; x 0 ; η n 1 ǫ. 1 When ne f the prcedures is asympttically better than the ther at a neighbrhd f x 0, we say δ 1 and δ are rdered at x 0. One may wnder if it is better t require the cnditin in 1 fr η n nly. The answer is n because it is pssible that L f,n ; x 0 ; η n is smaller than L f 1,n ; x 0 ; η n with high prbability fr ne sequence f η n yet the ppsite hlds fr a smaller sequence 0 < η n η n. This can happen, fr example, when δ 1 is asympttically better than δ glbally i.e., with n restrictin n η n but wrse lcally. In general, the space X can be decmpsed int three regins: thse x 0 at which δ 1 is asympttically better than δ, thse at which δ is asympttically better than δ 1, and the rest f x 0 at which δ 1 and δ cannt be cmpared accrding t Definitin 1. Lcal selectin can be hped t be successful nly fr the first tw regins. Definitin. A prcedure δ r { f n } n=1 is said t cnverge exactly at rate {a n } in prbability at η n - neighbrhd f x 0 if fr each nn-increasing sequence η n with 0 < η n η n, L f; x 0 ; η n = O p a n, and fr every 0 < ǫ < 1, there exists c ǫ > 0 such that when n is large enugh, P L f; x 0 ; η n c ǫ a n 1 ǫ. Definitin 3. A selectin methd is said t be lcally cnsistent in selectin at x 0 if δ 1 and δ are rdered at x 0 and the asympttically better prcedure at x 0 is selected with prbability ging t 1. If the selectin methd is lcally cnsistent at every x 0 at which the prcedures are rdered, it is said t be lcally cnsistent. In Definitin, the secnd cnditin simply says that the lss des nt cnverge faster than a n in an apprpriate sense. Fr the fllwing result, we assume that f 1,n and f,n cnverge exactly at rate {p n } and {q n } in prbability at η n -neighbrhd f x 0 respectively, where p n and q n are tw sequences f nn-increasing psitive numbers. 8

9 Sme technical cnditins are needed. Cnditin 1. There exists a sequence f psitive numbers A n such that fr j = 1,, f f j,n = O p A n. Cnditin. There exists η n > 0 such that δ 1 δ at x 0 ; η n r δ δ 1 at x 0 ; η n. Let j x 0 dente the better prcedure. Cnditin 3. There exists a sequence f psitive numbers {M n } such that fr j = 1,, fr each nn-increasing sequence 0 η n η n, we have f f j,n,x0;eη n f f = O p M n. j,n,x0;eη n Therem 1: Under Cnditins 1-3, as lng as n 1, n, and r n are chsen t satisfy 1: n r d n M n 1 ; n r d n maxp n 1, q n1 /1 A n1 ; 3 r n η n, we have that with prbability ging t 1, the better prcedure f j x 0,n will be selected. Frm the result, several factrs affect the ability f the L-CV t identify the best estimatr at a given x 0. The larger η n, the larger windw fr chsing the neighbr size r n fr L-CV. Intuitively, if η n is very clse t zer, which ccurs when the tw prcedures cnstantly switch psitin in terms f lcal perfrmance, it may nt be feasible t identify the lcally better ne at all. The cnstants M n and A n cme int the picture mre fr a technical reasn. Fr simplicity, fr the fllwing discussin, we assume that M n and A n are bth f rder 1. Then the requirements n the chice f data splitting rati and the neighbr size becme n r d n maxp n 1, q n1 and r n η n. Cnsider, fr example, η n = lg n 1. When at least ne f the prcedures is nnparametric with maxp n, q n cnverging mre slwly than the parametric rate n 1/, Yang 007 shwed that fr selecting the glbally better prcedure, it suffices t take n at least f the same rder as n 1 which is nt enugh fr cmparing parametric mdels as shwn in Sha With lcal selectin, hwever, we need t be mre careful s as t satisfy 1 n maxp n 1, q n 1 rd n lg n 1 d. If maxp n, q n is f rder n 1/3, then with n 1 = n, the cnditin is simply n 1/3d r n lg n 1 1. When the number f bservatins is mderately large, ne can use a data dependent apprach t chse r n. Fr example, taking r n t be the same fr all x, ne may cnsider anther level f data splitting fr empirically evaluating the perfrmance f each chice f r n. Theretical results are pssible, but we will nt pursue them in this wrk. 9

10 5 Preference regin selectin Let f 1 x = f 1,n1 x and f x = f,n1 x be the estimates by the tw regressin prcedures based n the first part f the data f size n 1. Fr a measurable set A, define fx; A = f 1 xi {x A} f xi {x/ A}. Since E fx; A fx = E f1 x fx I {x A} f x fx I {x/ A} = E f1 x fx I{x A} E f x fx I{x/ A}, it fllws that the preference regin A = {x : E f1 x fx E f x fx } is the best chice f A. Obviusly, we have E f; A f min E f 1 f, f f, i.e., if it is pssible t identify A, the estimatr fx; A is better r at least n wrse than the glbal winner between the tw candidate prcedures. The ideal lcal selectin can be arbitrarily better than the glbal selectin in terms f their risk rati. We cnsider a cllectin f sets f manageable cmplexity and try t find the best set A within the class. Let A n be a class f sets and let A n be the set that minimizes E f; A f ver A n. Then E f; A n f E f; A f is the apprximatin errr due t the use f a set in A n, which may nt cntain A. Of curse, A n is als unknwn, and needs t be estimated. Let µ B A dente the prbability under the distributin f X i f the symmetric difference f A and B. Assume that A n has metric entrpy bunded abve by H n ǫ under the distance da, B = µ B A fr cncept and related results invlving metric entrpy, see, e.g., Klmgrv and Tihmirv, 1959; Yatracs, 1985; Birgé, 1986; van de Geer, 1993; Yang and Barrn, Let A n,0 be a discretized cllectin f sets that serves as an ǫ n -net fr A n under d. Let Ân be the minimizer f the empirical predictin errr ver A n,0, i.e., 1 Â n = arg min A A n,0 n n i=n 11 Y i fx i ; A. Therem : Assume that the errrs are nrmally distributed with mean zer and variance σ > 0, and f j,n f C a.s. fr sme cnstant C < fr bth prcedures. Then the final estimatr 10

11 f; Ân satisfies E f; Ân f C 1 E f; A n f ǫ n H nǫ n n C E f; A f E f; A n f; A ǫ n H nǫ n, n where the cnstants C 1 and C depend nly n C and σ. T ptimize the risk bund abve, with A n given, we need t balance Hnǫn n and ǫ n. The issue becmes mre cmplicated when ne needs t chse A n. Clearly, a large chice f A n reduces the ptential apprximatin errr but at the same time increases the estimatin errr due t searching ver a larger class f sets A. One apprach t handling the apprximatin errr is t assume that A is in a given cllectin f sets B, and then characterize the unifrm apprximatin errr f sets in B by sets in A n. Defined γ n = sup B B inf A An µ B A. Under prper assumptins n B and A n, the rate f cnvergence f γ n can be derived and then used fr btaining the cnvergence rate f the final estimatr frm the lcalized selectin. An example Let A k cnsist f all pssible unins f at mst k cubes in X = [0, 1] d. Let B be the cllectin f all the sets that each can be well apprximated by a set in A k in the sense that fr each B B, there exists a set A in A k such that µ B A ck τ fr sme cnstants c, τ > 0. Fr simplicity, we assume that X i has a unifrm distributin n [0, 1] d, althugh a similar result hlds if X i has a unifrmly upper bunded Lebesgue density. T btain an ε-net in A k fr B, we first chse k = k ε such that ck τ ε/ k ε is then f rder ε 1/τ. Then fr any B B, there exists a set A in A kε such that µ B A ε/. Cnsequently, an ε/-net in A kε will be an ε-net fr B. Nw if we have a grid n [0, 1] d f width ε/ k fr each crdinate, then the cllectin f all cubes with vertices n the grid frm an ε/ k-net fr A kε and thus als an ε-net fr B. The number f pssible unins f k such cubes is f rder kεd kε = ε 11/τε 1/τd. ε It fllws that the metric entrpy f B is f rder O ε 1/τ lgε 1. If we select A ver the ε-net in A k by crss validatin with a prper discretizatin, by Therem, with a chice f n 1 and n bth f rder n, the cmbined estimatr satisfies E f; Ân f = O E f; A f ǫ n lg ǫ 1 n. nǫ 1/τ n 11

12 Balancing the last tw terms in the abve expressin, the ptimal chice f ǫ n is lg n/n τ τ1, and cnsequently the risk bund becmes E f; Ân f = O E f; A f lg n/n τ τ1. Thus, when τ is large, the additinal term lg n/n τ τ1 is clse t lg n/n, which is typically negligible fr nnparametric regressin. Then we achieve the perfrmance f the ideal lcalized selectin up t a relatively small discrepancy term. 6 Cncluding remarks Overall perfrmance f a candidate mdel prcedure has been the dminating measure used fr mdel selectin. If ne candidate mdel is thught t be true r ptimal, it is then ratinal t try t identify it. Hwever, in many applicatins, this practice is sub-ptimal because the glbally best candidate prcedure, even if it is much better than thers, may still have unsatisfactry lcal behavirs in certain regins, which can be well remedied with helps frm ther candidates that are glbally inferir. We tk tw directins in lcalized mdel selectin: a lcalized crss validatin that selects a mdel/prcedure at each x value based n perfrmance assessment in a neighbrhd f x, and a preference regin selectin frm a cllectin f sets, which tries t estimate the regins where each candidate perfrms better. Fr the lcalized crss validatin, as lng as the neighbrhd size and the data splitting rati are chsen prperly, the lcally better estimatr will be selected with high prbability. Fr preference regin selectin, when the cmplexity f the cllectin f the candidate sets f the preference regin is prperly cntrlled, the final prcedure based n the empirically selected preference regin behaves well in risk. Besides glbal selectin and lcalized selectin f a mdel, smetimes it is advantageus t cnsider glbal cmbinatin f the candidates with weights glbally determined, r lcalized cmbinatin where weights depend n the x value see, e.g., Pan, Xia and Huang, 006. Fr high-dimensinal r cmplex data, these alternatives can prvide flexibility needed t further imprve perfrmance f the candidate prcedures. 7 Appendix Prf f Therem 1. Much f the prf fllws frm Yang 007, except that lcalized selectin is cnsidered in this wrk. We first fcus n the analysis withut multiple data splittings. Withut lss f generality, assume that δ 1 is asympttically better than δ at x 0. Let I = I x0,r n = {i : n 1 1 i n 1

13 and X i Bx 0 ; r n } dente the bservatins in the evaluatin set with X i clse t x 0, and let ñ be the size f I. Because LCV f j,n1 = Y i f j,n1 X i fx i f j,n1 X i ε i = = ε i LCV f 1,n1 LCV f,n1 is equivalent t fx i f j,n1 X i ε i f,n1 X i f 1,n1 X i fx i f,n1 X i Cnditinal n Z 1 = X i, Y i n1 i=1 and X = X n11,..., X n, assuming larger than fx i f 1,n1 X i, by Chebyshev s inequality, we have ε i fx i f j,n1 X i, fx i f 1,n1 X i. fx i f,n1 X i is P LCV f 1,n1 > LCV f,n1 Z 1, X σ min 1, f,n1 X i f 1,n1 X i fx i f,n1 X i fx i f. 1,n1 X i Let Q n dente the rati in the upper bund in the abve inequality and let S n be the event f Then because fx i f,n1 X i > fx i f 1,n1 X i. P LCV f 1,n1 > LCV f,n1 { = P LCV f 1,n1 > LCV f } {,n1 S n P CV f 1,n1 > CV f },n1 Sn c E P CV f 1,n1 > CV f,n1 Z 1, X I Sn PSn c E min 1, Q n PS c n, fr cnsistency, it suffices t shw PS c n 0 and Q n 0 in prbability. Suppse we can shw that fr each ǫ > 0, there exists α ǫ > 0 such that when n is large enugh, fx i f,n1 X i P fx i f 1 α ǫ 1 ǫ. 1,n1 X i Then PS n 1 ǫ and thus PS c n 0 as n. Since f,n1 X i f 1,n1 X i fx i f 1,n1 X i 13 fx i f,n1 X i,

14 with prbability n less than 1 ǫ, we have Q n 8σ fx i f 1,n1 X i fx i f,n1 X i 1 1 1α ǫ fx i f,n1 X i 8σ α ǫ 1α ǫ fx i f,n1 X i. 3 Frm and 3, t shw PS c n 0 and Q n 0 in prbability, it suffices t shw and fx i f,n1 X i in prbability. Suppse a slight relaxatin f Cnditin 1 hlds: fr every ǫ > 0, there exists A n1,ǫ such that fr j = 1,, when n 1 is large enugh, f P fj,n1 A n1,ǫ ǫ. { f Let H n1 dente the event max f1,n1, f f,n1 A n1,ǫ fx i f f j,n1 X i fj,n1 Cnditinal n Z 1 and H n1, V ar Z 1W n11 E Z 1 }. Then n H n1, we have W i = with X i Bx 0 ; r n is bunded between A n1,ǫ and A n1,ǫ. fx n11 f f j,n1 X n11 = fj,n1,x 0;r n, where the subscript Z 1 in V ar Z 1 and E Z 1 is used t dente the cnditinal expectatin given Z 1. Thus cnditinal n Z 1, n H n1, by Bernstein s inequality see, e.g., Pllard 198, page 193, fr each x > 0, P Z 1 fx i f f 1,n1 X i ñ f1,n1 x,x 0;r n Taking x = β n ñ f f1,n1 exp 1 f ñ f1,n1 P Z 1 x An1,ǫ x,x 3 0;r n., the abve inequality becmes,x 0;r n fx i f f 1,n1 X i 1 βn ñ f1,n1,x 0;r n exp 1 f f 1,n1 β nñ f f 1,n1 An1,ǫ β n 3,x 0;r n,x 0;r n f f 1,n1,x 0;r n Under Cnditin, fr every ǫ > 0, there exists α ǫ > 0 such that when n is large enugh, f f,n1,x P 0;r n f f 1 αǫ ǫ. 1,n1,x 0;r n 1.

15 Take β n such that 1 β n = f f,n1,x 0;r n 1 α ǫ/ f f 1,n1,x 0;r n. Then with prbability at least 1 ǫ, β n α ǫ / 1α ǫ /. Let D n dente this event. Then n D n, we have and P Z 1 = P Z 1 P Z 1 If we have β n = = f f,n1 1 α f ǫ / f1,n1,x 0;r n,x 0;r n f 1 α ǫ / f1,n1,x 0;r n f f 1α ǫ,n1 / f f,n1 b,x 0 ;rn,x 1α 0;r n ǫ f 1 α ǫ / f1,n1,x 0;r n α ǫ f f,n1,x 0;r n f 1 α ǫ 1 α ǫ / f1,n1,x 0;r n, fx i f f 1,n1 X i 1 βn ñ f1,n1,x 0;r n fx i f ñ f 1,n1 X i 1 α ǫ / f,n1,x 0;r n fx i f α ǫ f f,n1,x 1,n1 X i 1 0;r n 1 α ǫ1 α ǫ/ f f 1,n1 α exp ǫ 8 1 α ǫ 1 α ǫ / f f 1,n1,x 0;r n f ñ f,n1 f f 1,n1,x 0;r n ñ f f,n1,x 0;r n,x 0;r n α ǫa n1,ǫ f 31α ǫ1α / f,n1 ǫ ñ f f1,n1,x 0;r n.,x 0;r n in prbability, 5 f ñ f,n1,x 0;r n A n1,ǫ in prbability, 6 then the upper bund in the last inequality abve cnverges t zer in prbability. Frm these pieces, we can cnclude that P fx i f ñ f 1,n1 X i 1 α ǫ / f,n1 3ǫ ǫ, n, 7,x 0;r n,x 0;r n 15

16 fr sme ǫ, n 0 as n. Indeed, fr every given ǫ > 0, when n is large enugh, 1 P fx i ñ f 1 f 1,n1 X i 1 α ǫ / f,n1,x 0;r n { PHn c 1 PDn c 1 P H n1 D n fx i ñ f 1 1,n1 X i f 1 α ǫ/ f,n1 3ǫ EP H n1 D n { 1 fx i ñ f 1 f 1,n1 X i 1 α ǫ / f,n1 α 3ǫ E exp ǫ 8 1 α ǫ 1 α ǫ/ 3ǫ ǫ, n, f f 1,n1,x 0;r n ñ f f,n1,x 0;r n,x 0;r n } Z 1 α ǫa n1,ǫ f 31α ǫ1α / f,n1 ǫ,x 0;r n,x 0;r n } where the expectatin in the upper bund f the last inequality abve i.e., ǫ, n cnverges t zer due t the cnvergence in prbability t zer f the randm variables f the expnential expressin and their unifrm integrability since they are bunded abve by 1, prvided that 5 and 6 hld. The assertin f 7 then fllws. Fr the ther estimatr, similarly, fr 0 < β n < 1, we have P Z 1 fx i f f,n1 X i 1 βn ñ f,n1,x 0;r n f ñ β n f,n1 exp 1,x 0;r n f f,n1 f f.,n1 An1,ǫ βn e,x 3 0;r n,x 0;r n If we have f ñ β n f,n1,x 0;r n f f in prbability, 8,n1,x 0;r n f ñ βn f,n1,x 0;r n A n1,ǫ in prbability, 9 then fllwing a similar argument used fr f 1,n1, we have P fx i f f,n1 X i 1 βn ñ f,n ,x 0;r n f Frm this, if ñ f,n1 in prbability and β n is bunded away frm 1, then hlds.,x 0;r n If in additin, we can chse β n 0, then fr each given ǫ, we have 1 β n f f,n1,x 0;r n > 1α ǫ f 1α / f,n1 fr sme small α ǫ > 0 when n 1 is large enugh. Nw fr every ǫ > 0, we can ǫ,x 0;r n 16

17 find ǫ > 0 such that 3ǫ ǫ/3 and there exists an integer n 0 such that when n n 0 the prbability in 10 is upper bunded by ǫ/3 and ǫ, n ǫ/3. Cnsequently when n n 0, fx i f,n1 X i P fx i f 1 α ǫ 1 ǫ. 1,n1 X i Recall that we needed the cnditins 5, 6, 8, and 9 fr t hld. Under Cnditin 3, en f b f,n1,x 0 ;rn f f b is lwer bunded in rder in prbability by en f f b,n1,x 0 ;rn 1,n1 M,x n 0 ;rn 1 f f b. Frm all abve, since 1,n1,x 0 ;rn f 1,n1 and f,n1 cnverge exactly at rates p n and q n respectively under the L lss, we knw that fr the cnclusin f Therem 1 t hld, it suffices t have the requirements: fr every ǫ > 0, fr sme β n 0, ñ β n M n 1, ñ q n1 /p n1 M n 1, ñ βn q n 1 A n1,ǫ, ñ q n 1 A n1,ǫ, ñ q n 1. Nte that ñ is a randm variable that has a binmial distributin with n trials and success prbability P X i Bx 0 ; r n. Under the assumptin that the density f X is bunded away frm zer in a neighbrhd f x 0, P X i Bx 0 ; r n is f rder r d n. We need t lwer bund ñ in prbability. Fr 1 j n, let D j be independent Bernulli randm variables with success prbability β. Then applying the Bernstein s inequality see, e.g., Pllard 198, p. 193, we have n P D j n β/ exp 3n β j=1 Fr ur setting, β is f rder r d n. Thus ñ is at least f rder n r d n in prbability if n r d n. Cnsequently, the earlier requirements n the splitting rati becme that fr every ǫ > 0, fr sme β n 0, n r d n β n M n 1, n r d n q n1 /p n1 M n 1, n r d n β n q n 1 A n1,ǫ, n r d n q n 1 A n1,ǫ, n r d n q n 1. 17

18 f Under Cnditin 1, fr every ǫ > 0, there exists a cnstant B ǫ > 0 such that P fj,n1 B ǫ A n1 ǫ when n 1 is large enugh. That is, fr a given ǫ > 0, we can take A n1,ǫ = OA n1. Therefre if we have n r d n M n 1 and n r d n q n 1 / 1 A n1, then we can find β n 0 such that the abve 5 displayed requirements are all satisfied. Let π dente a permutatin f the bservatins, and let Π dente a set f such permutatins. Let LCV π f j,n1 be the L-CV criterin value under the permutatin π. If LCV π f 1,n1 LCV π f,n1, then let τ π = 1 and therwise let τ π = 0. Let W dente the values f X 1, Y 1,..., X n, Y n ignring the rders. Under the i.i.d assumptin n the bservatins, bviusly, cnditinal n W, every rdering f these values has exactly the same prbability and thus PLCV f 1,n1 LCV f,n1 = EPLCV f 1,n1 LCV f,n1 W π Π = E τ π. Π Frm the earlier analysis, we knw that PLCV f 1,n1 LCV f,n1 1. Thus E π Π τ π/ Π 1. Since π Π τ π/ Π is between 0 and 1, fr its expectatin t cnverge t 1, we must have π Π τ π/ Π 1 in prbability. This cmpletes the prf f Therem 1. Prf f Therem : Let à n be the minimizer f E f; A f ver A A n,0. By the results f Wegkamp 003, Sectin, in particular, Therem.1, we have E f; Ân f E f; à n f 1n 6C lg A n,0 e 1/6C 16σ lg A n,0 e 1/16σ n E f; à n f B 1 B H n ǫ n, n n where B 1 and B are cnstants that depend nly n C and σ. Nte that fx; A fx; B dµ f1 x fx I {x A} I {x B} dµ f x fx I {x A} I {x B} dµ = f1 x fx dµ f x fx dµ = B A B A B c A c f1 x fx dµ f x fx dµ C µ B A. Tgether with that A n,0 is an ǫ n -net under d, we have E f; à n f E f; A n f 8C ǫ n. The cnclusins then fllw. This cmpletes the prf f Therem. 18

19 8 Acknwledgments This research was supprted by US Natinal Science Fundatin CAREER Grant DMS We thank three referees and the editrs fr helpful cmments n imprving the paper. References [1] Allen, D.M. 197 The relatinship between variable selectin and data augmentatin and a methd fr predictin. Technmetrics, 16, [] Birgé, L On estimating a density using Hellinger distance and sme ther strange facts. Prbability Thery and Related Fields, 71, [3] Burman, P A cmparative study f rdinary crss-validatin, ν-fld crss-validatin and the repeated learning-testing methds, Bimetrika, 76, [] Geisser, S The predictive sample reuse methd with applicatins, Jurnal f the American Statistical Assciatin, 70, [5] Klmgrv, A.N. and Tihmirv, V.M ǫ-entrpy and ǫ-capacity f sets in functin spaces. Uspehi Mat. Nauk 1, [6] Li, K.-C Asympttic ptimality fr C p, C L, crss-validatin and generalized crss-validatin: Discrete index set, The Annals f Statistics, 15, [7] Pan, W., Xia, G. and Huang, X. 006 Using input dependent weights fr mdel cmbinatin and mdel selectin with multiple surces f data, Statistics Sinica, 16, [8] Pllard, D. 198 Cnvergence f Stchastic Prcesses. Springer, New Yrk. [9] Sha, Jun 1993 Linear mdel selectin by crss-validatin Jurnal f the American Statistical Assciatin, 88, [10] Sha, J An asympttic thery fr linear mdel selectin with discussin Statistica Sinica, 7, 1-. [11] Stne, M. 197 Crss-validatin chice and assessment f statistical predictins. Jurnal f the Ryal Statistical Sciety, Ser.B, 36, [1] van de Geer, S Hellinger-cnsistency f certain nnparametric maximum likelihd estimatrs. Annals f Statistics, 1, 1-. [13] Wegkamp, M.H Mdel selectin in nnparametric regressin. Ann. Statist., 31, [1] Yatracs, Y.G Rates f cnvergence f minimum distance estimatrs and Klmgrv s entrpy. Annals f Statistics, 13, [15] Yang, Y. 007 Cnsistency f crss validatin fr cmparing regressin prcedures. Accepted by Ann. Statistics. [16] Yang, Y. and Barrn, A.R Infrmatin-theretic determinatin f minimax rates f cnvergence. Ann. Statistics, 7, [17] Zhang, P Mdel selectin via multifld crss validatin, Annals f Statistics, 1,

Resampling Methods. Chapter 5. Chapter 5 1 / 52

Resampling Methods. Chapter 5. Chapter 5 1 / 52 Resampling Methds Chapter 5 Chapter 5 1 / 52 1 51 Validatin set apprach 2 52 Crss validatin 3 53 Btstrap Chapter 5 2 / 52 Abut Resampling An imprtant statistical tl Pretending the data as ppulatin and

More information

What is Statistical Learning?

What is Statistical Learning? What is Statistical Learning? Sales 5 10 15 20 25 Sales 5 10 15 20 25 Sales 5 10 15 20 25 0 50 100 200 300 TV 0 10 20 30 40 50 Radi 0 20 40 60 80 100 Newspaper Shwn are Sales vs TV, Radi and Newspaper,

More information

k-nearest Neighbor How to choose k Average of k points more reliable when: Large k: noise in attributes +o o noise in class labels

k-nearest Neighbor How to choose k Average of k points more reliable when: Large k: noise in attributes +o o noise in class labels Mtivating Example Memry-Based Learning Instance-Based Learning K-earest eighbr Inductive Assumptin Similar inputs map t similar utputs If nt true => learning is impssible If true => learning reduces t

More information

Bootstrap Method > # Purpose: understand how bootstrap method works > obs=c(11.96, 5.03, 67.40, 16.07, 31.50, 7.73, 11.10, 22.38) > n=length(obs) >

Bootstrap Method > # Purpose: understand how bootstrap method works > obs=c(11.96, 5.03, 67.40, 16.07, 31.50, 7.73, 11.10, 22.38) > n=length(obs) > Btstrap Methd > # Purpse: understand hw btstrap methd wrks > bs=c(11.96, 5.03, 67.40, 16.07, 31.50, 7.73, 11.10, 22.38) > n=length(bs) > mean(bs) [1] 21.64625 > # estimate f lambda > lambda = 1/mean(bs);

More information

Distributions, spatial statistics and a Bayesian perspective

Distributions, spatial statistics and a Bayesian perspective Distributins, spatial statistics and a Bayesian perspective Dug Nychka Natinal Center fr Atmspheric Research Distributins and densities Cnditinal distributins and Bayes Thm Bivariate nrmal Spatial statistics

More information

Tree Structured Classifier

Tree Structured Classifier Tree Structured Classifier Reference: Classificatin and Regressin Trees by L. Breiman, J. H. Friedman, R. A. Olshen, and C. J. Stne, Chapman & Hall, 98. A Medical Eample (CART): Predict high risk patients

More information

Lecture 2: Supervised vs. unsupervised learning, bias-variance tradeoff

Lecture 2: Supervised vs. unsupervised learning, bias-variance tradeoff Lecture 2: Supervised vs. unsupervised learning, bias-variance tradeff Reading: Chapter 2 STATS 202: Data mining and analysis September 27, 2017 1 / 20 Supervised vs. unsupervised learning In unsupervised

More information

Lecture 2: Supervised vs. unsupervised learning, bias-variance tradeoff

Lecture 2: Supervised vs. unsupervised learning, bias-variance tradeoff Lecture 2: Supervised vs. unsupervised learning, bias-variance tradeff Reading: Chapter 2 STATS 202: Data mining and analysis September 27, 2017 1 / 20 Supervised vs. unsupervised learning In unsupervised

More information

Resampling Methods. Cross-validation, Bootstrapping. Marek Petrik 2/21/2017

Resampling Methods. Cross-validation, Bootstrapping. Marek Petrik 2/21/2017 Resampling Methds Crss-validatin, Btstrapping Marek Petrik 2/21/2017 Sme f the figures in this presentatin are taken frm An Intrductin t Statistical Learning, with applicatins in R (Springer, 2013) with

More information

Lyapunov Stability Stability of Equilibrium Points

Lyapunov Stability Stability of Equilibrium Points Lyapunv Stability Stability f Equilibrium Pints 1. Stability f Equilibrium Pints - Definitins In this sectin we cnsider n-th rder nnlinear time varying cntinuus time (C) systems f the frm x = f ( t, x),

More information

COMP 551 Applied Machine Learning Lecture 5: Generative models for linear classification

COMP 551 Applied Machine Learning Lecture 5: Generative models for linear classification COMP 551 Applied Machine Learning Lecture 5: Generative mdels fr linear classificatin Instructr: Herke van Hf (herke.vanhf@mail.mcgill.ca) Slides mstly by: Jelle Pineau Class web page: www.cs.mcgill.ca/~hvanh2/cmp551

More information

Pattern Recognition 2014 Support Vector Machines

Pattern Recognition 2014 Support Vector Machines Pattern Recgnitin 2014 Supprt Vectr Machines Ad Feelders Universiteit Utrecht Ad Feelders ( Universiteit Utrecht ) Pattern Recgnitin 1 / 55 Overview 1 Separable Case 2 Kernel Functins 3 Allwing Errrs (Sft

More information

3.4 Shrinkage Methods Prostate Cancer Data Example (Continued) Ridge Regression

3.4 Shrinkage Methods Prostate Cancer Data Example (Continued) Ridge Regression 3.3.4 Prstate Cancer Data Example (Cntinued) 3.4 Shrinkage Methds 61 Table 3.3 shws the cefficients frm a number f different selectin and shrinkage methds. They are best-subset selectin using an all-subsets

More information

Hypothesis Tests for One Population Mean

Hypothesis Tests for One Population Mean Hypthesis Tests fr One Ppulatin Mean Chapter 9 Ala Abdelbaki Objective Objective: T estimate the value f ne ppulatin mean Inferential statistics using statistics in rder t estimate parameters We will be

More information

Perfrmance f Sensitizing Rules n Shewhart Cntrl Charts with Autcrrelated Data Key Wrds: Autregressive, Mving Average, Runs Tests, Shewhart Cntrl Chart

Perfrmance f Sensitizing Rules n Shewhart Cntrl Charts with Autcrrelated Data Key Wrds: Autregressive, Mving Average, Runs Tests, Shewhart Cntrl Chart Perfrmance f Sensitizing Rules n Shewhart Cntrl Charts with Autcrrelated Data Sandy D. Balkin Dennis K. J. Lin y Pennsylvania State University, University Park, PA 16802 Sandy Balkin is a graduate student

More information

A Matrix Representation of Panel Data

A Matrix Representation of Panel Data web Extensin 6 Appendix 6.A A Matrix Representatin f Panel Data Panel data mdels cme in tw brad varieties, distinct intercept DGPs and errr cmpnent DGPs. his appendix presents matrix algebra representatins

More information

CS 477/677 Analysis of Algorithms Fall 2007 Dr. George Bebis Course Project Due Date: 11/29/2007

CS 477/677 Analysis of Algorithms Fall 2007 Dr. George Bebis Course Project Due Date: 11/29/2007 CS 477/677 Analysis f Algrithms Fall 2007 Dr. Gerge Bebis Curse Prject Due Date: 11/29/2007 Part1: Cmparisn f Srting Algrithms (70% f the prject grade) The bjective f the first part f the assignment is

More information

, which yields. where z1. and z2

, which yields. where z1. and z2 The Gaussian r Nrmal PDF, Page 1 The Gaussian r Nrmal Prbability Density Functin Authr: Jhn M Cimbala, Penn State University Latest revisin: 11 September 13 The Gaussian r Nrmal Prbability Density Functin

More information

CAUSAL INFERENCE. Technical Track Session I. Phillippe Leite. The World Bank

CAUSAL INFERENCE. Technical Track Session I. Phillippe Leite. The World Bank CAUSAL INFERENCE Technical Track Sessin I Phillippe Leite The Wrld Bank These slides were develped by Christel Vermeersch and mdified by Phillippe Leite fr the purpse f this wrkshp Plicy questins are causal

More information

Eric Klein and Ning Sa

Eric Klein and Ning Sa Week 12. Statistical Appraches t Netwrks: p1 and p* Wasserman and Faust Chapter 15: Statistical Analysis f Single Relatinal Netwrks There are fur tasks in psitinal analysis: 1) Define Equivalence 2) Measure

More information

A New Evaluation Measure. J. Joiner and L. Werner. The problems of evaluation and the needed criteria of evaluation

A New Evaluation Measure. J. Joiner and L. Werner. The problems of evaluation and the needed criteria of evaluation III-l III. A New Evaluatin Measure J. Jiner and L. Werner Abstract The prblems f evaluatin and the needed criteria f evaluatin measures in the SMART system f infrmatin retrieval are reviewed and discussed.

More information

7 TH GRADE MATH STANDARDS

7 TH GRADE MATH STANDARDS ALGEBRA STANDARDS Gal 1: Students will use the language f algebra t explre, describe, represent, and analyze number expressins and relatins 7 TH GRADE MATH STANDARDS 7.M.1.1: (Cmprehensin) Select, use,

More information

Part 3 Introduction to statistical classification techniques

Part 3 Introduction to statistical classification techniques Part 3 Intrductin t statistical classificatin techniques Machine Learning, Part 3, March 07 Fabi Rli Preamble ØIn Part we have seen that if we knw: Psterir prbabilities P(ω i / ) Or the equivalent terms

More information

On Huntsberger Type Shrinkage Estimator for the Mean of Normal Distribution ABSTRACT INTRODUCTION

On Huntsberger Type Shrinkage Estimator for the Mean of Normal Distribution ABSTRACT INTRODUCTION Malaysian Jurnal f Mathematical Sciences 4(): 7-4 () On Huntsberger Type Shrinkage Estimatr fr the Mean f Nrmal Distributin Department f Mathematical and Physical Sciences, University f Nizwa, Sultanate

More information

MATCHING TECHNIQUES. Technical Track Session VI. Emanuela Galasso. The World Bank

MATCHING TECHNIQUES. Technical Track Session VI. Emanuela Galasso. The World Bank MATCHING TECHNIQUES Technical Track Sessin VI Emanuela Galass The Wrld Bank These slides were develped by Christel Vermeersch and mdified by Emanuela Galass fr the purpse f this wrkshp When can we use

More information

COMP 551 Applied Machine Learning Lecture 4: Linear classification

COMP 551 Applied Machine Learning Lecture 4: Linear classification COMP 551 Applied Machine Learning Lecture 4: Linear classificatin Instructr: Jelle Pineau (jpineau@cs.mcgill.ca) Class web page: www.cs.mcgill.ca/~jpineau/cmp551 Unless therwise nted, all material psted

More information

Chapter 3: Cluster Analysis

Chapter 3: Cluster Analysis Chapter 3: Cluster Analysis } 3.1 Basic Cncepts f Clustering 3.1.1 Cluster Analysis 3.1. Clustering Categries } 3. Partitining Methds 3..1 The principle 3.. K-Means Methd 3..3 K-Medids Methd 3..4 CLARA

More information

MATCHING TECHNIQUES Technical Track Session VI Céline Ferré The World Bank

MATCHING TECHNIQUES Technical Track Session VI Céline Ferré The World Bank MATCHING TECHNIQUES Technical Track Sessin VI Céline Ferré The Wrld Bank When can we use matching? What if the assignment t the treatment is nt dne randmly r based n an eligibility index, but n the basis

More information

Internal vs. external validity. External validity. This section is based on Stock and Watson s Chapter 9.

Internal vs. external validity. External validity. This section is based on Stock and Watson s Chapter 9. Sectin 7 Mdel Assessment This sectin is based n Stck and Watsn s Chapter 9. Internal vs. external validity Internal validity refers t whether the analysis is valid fr the ppulatin and sample being studied.

More information

Least Squares Optimal Filtering with Multirate Observations

Least Squares Optimal Filtering with Multirate Observations Prc. 36th Asilmar Cnf. n Signals, Systems, and Cmputers, Pacific Grve, CA, Nvember 2002 Least Squares Optimal Filtering with Multirate Observatins Charles W. herrien and Anthny H. Hawes Department f Electrical

More information

x 1 Outline IAML: Logistic Regression Decision Boundaries Example Data

x 1 Outline IAML: Logistic Regression Decision Boundaries Example Data Outline IAML: Lgistic Regressin Charles Suttn and Victr Lavrenk Schl f Infrmatics Semester Lgistic functin Lgistic regressin Learning lgistic regressin Optimizatin The pwer f nn-linear basis functins Least-squares

More information

PSU GISPOPSCI June 2011 Ordinary Least Squares & Spatial Linear Regression in GeoDa

PSU GISPOPSCI June 2011 Ordinary Least Squares & Spatial Linear Regression in GeoDa There are tw parts t this lab. The first is intended t demnstrate hw t request and interpret the spatial diagnstics f a standard OLS regressin mdel using GeDa. The diagnstics prvide infrmatin abut the

More information

Module 4: General Formulation of Electric Circuit Theory

Module 4: General Formulation of Electric Circuit Theory Mdule 4: General Frmulatin f Electric Circuit Thery 4. General Frmulatin f Electric Circuit Thery All electrmagnetic phenmena are described at a fundamental level by Maxwell's equatins and the assciated

More information

The blessing of dimensionality for kernel methods

The blessing of dimensionality for kernel methods fr kernel methds Building classifiers in high dimensinal space Pierre Dupnt Pierre.Dupnt@ucluvain.be Classifiers define decisin surfaces in sme feature space where the data is either initially represented

More information

COMP 551 Applied Machine Learning Lecture 9: Support Vector Machines (cont d)

COMP 551 Applied Machine Learning Lecture 9: Support Vector Machines (cont d) COMP 551 Applied Machine Learning Lecture 9: Supprt Vectr Machines (cnt d) Instructr: Herke van Hf (herke.vanhf@mail.mcgill.ca) Slides mstly by: Class web page: www.cs.mcgill.ca/~hvanh2/cmp551 Unless therwise

More information

Particle Size Distributions from SANS Data Using the Maximum Entropy Method. By J. A. POTTON, G. J. DANIELL AND B. D. RAINFORD

Particle Size Distributions from SANS Data Using the Maximum Entropy Method. By J. A. POTTON, G. J. DANIELL AND B. D. RAINFORD 3 J. Appl. Cryst. (1988). 21,3-8 Particle Size Distributins frm SANS Data Using the Maximum Entrpy Methd By J. A. PTTN, G. J. DANIELL AND B. D. RAINFRD Physics Department, The University, Suthamptn S9

More information

Physics 2B Chapter 23 Notes - Faraday s Law & Inductors Spring 2018

Physics 2B Chapter 23 Notes - Faraday s Law & Inductors Spring 2018 Michael Faraday lived in the Lndn area frm 1791 t 1867. He was 29 years ld when Hand Oersted, in 1820, accidentally discvered that electric current creates magnetic field. Thrugh empirical bservatin and

More information

IN a recent article, Geary [1972] discussed the merit of taking first differences

IN a recent article, Geary [1972] discussed the merit of taking first differences The Efficiency f Taking First Differences in Regressin Analysis: A Nte J. A. TILLMAN IN a recent article, Geary [1972] discussed the merit f taking first differences t deal with the prblems that trends

More information

Modelling of Clock Behaviour. Don Percival. Applied Physics Laboratory University of Washington Seattle, Washington, USA

Modelling of Clock Behaviour. Don Percival. Applied Physics Laboratory University of Washington Seattle, Washington, USA Mdelling f Clck Behaviur Dn Percival Applied Physics Labratry University f Washingtn Seattle, Washingtn, USA verheads and paper fr talk available at http://faculty.washingtn.edu/dbp/talks.html 1 Overview

More information

Admissibility Conditions and Asymptotic Behavior of Strongly Regular Graphs

Admissibility Conditions and Asymptotic Behavior of Strongly Regular Graphs Admissibility Cnditins and Asympttic Behavir f Strngly Regular Graphs VASCO MOÇO MANO Department f Mathematics University f Prt Oprt PORTUGAL vascmcman@gmailcm LUÍS ANTÓNIO DE ALMEIDA VIEIRA Department

More information

SUPPLEMENTARY MATERIAL GaGa: a simple and flexible hierarchical model for microarray data analysis

SUPPLEMENTARY MATERIAL GaGa: a simple and flexible hierarchical model for microarray data analysis SUPPLEMENTARY MATERIAL GaGa: a simple and flexible hierarchical mdel fr micrarray data analysis David Rssell Department f Bistatistics M.D. Andersn Cancer Center, Hustn, TX 77030, USA rsselldavid@gmail.cm

More information

1996 Engineering Systems Design and Analysis Conference, Montpellier, France, July 1-4, 1996, Vol. 7, pp

1996 Engineering Systems Design and Analysis Conference, Montpellier, France, July 1-4, 1996, Vol. 7, pp THE POWER AND LIMIT OF NEURAL NETWORKS T. Y. Lin Department f Mathematics and Cmputer Science San Jse State University San Jse, Califrnia 959-003 tylin@cs.ssu.edu and Bereley Initiative in Sft Cmputing*

More information

Support-Vector Machines

Support-Vector Machines Supprt-Vectr Machines Intrductin Supprt vectr machine is a linear machine with sme very nice prperties. Haykin chapter 6. See Alpaydin chapter 13 fr similar cntent. Nte: Part f this lecture drew material

More information

COMP 551 Applied Machine Learning Lecture 11: Support Vector Machines

COMP 551 Applied Machine Learning Lecture 11: Support Vector Machines COMP 551 Applied Machine Learning Lecture 11: Supprt Vectr Machines Instructr: (jpineau@cs.mcgill.ca) Class web page: www.cs.mcgill.ca/~jpineau/cmp551 Unless therwise nted, all material psted fr this curse

More information

Department of Economics, University of California, Davis Ecn 200C Micro Theory Professor Giacomo Bonanno. Insurance Markets

Department of Economics, University of California, Davis Ecn 200C Micro Theory Professor Giacomo Bonanno. Insurance Markets Department f Ecnmics, University f alifrnia, Davis Ecn 200 Micr Thery Prfessr Giacm Bnann Insurance Markets nsider an individual wh has an initial wealth f. ith sme prbability p he faces a lss f x (0

More information

initially lcated away frm the data set never win the cmpetitin, resulting in a nnptimal nal cdebk, [2] [3] [4] and [5]. Khnen's Self Organizing Featur

initially lcated away frm the data set never win the cmpetitin, resulting in a nnptimal nal cdebk, [2] [3] [4] and [5]. Khnen's Self Organizing Featur Cdewrd Distributin fr Frequency Sensitive Cmpetitive Learning with One Dimensinal Input Data Aristides S. Galanpuls and Stanley C. Ahalt Department f Electrical Engineering The Ohi State University Abstract

More information

4th Indian Institute of Astrophysics - PennState Astrostatistics School July, 2013 Vainu Bappu Observatory, Kavalur. Correlation and Regression

4th Indian Institute of Astrophysics - PennState Astrostatistics School July, 2013 Vainu Bappu Observatory, Kavalur. Correlation and Regression 4th Indian Institute f Astrphysics - PennState Astrstatistics Schl July, 2013 Vainu Bappu Observatry, Kavalur Crrelatin and Regressin Rahul Ry Indian Statistical Institute, Delhi. Crrelatin Cnsider a tw

More information

CHAPTER 24: INFERENCE IN REGRESSION. Chapter 24: Make inferences about the population from which the sample data came.

CHAPTER 24: INFERENCE IN REGRESSION. Chapter 24: Make inferences about the population from which the sample data came. MATH 1342 Ch. 24 April 25 and 27, 2013 Page 1 f 5 CHAPTER 24: INFERENCE IN REGRESSION Chapters 4 and 5: Relatinships between tw quantitative variables. Be able t Make a graph (scatterplt) Summarize the

More information

Midwest Big Data Summer School: Machine Learning I: Introduction. Kris De Brabanter

Midwest Big Data Summer School: Machine Learning I: Introduction. Kris De Brabanter Midwest Big Data Summer Schl: Machine Learning I: Intrductin Kris De Brabanter kbrabant@iastate.edu Iwa State University Department f Statistics Department f Cmputer Science June 24, 2016 1/24 Outline

More information

CN700 Additive Models and Trees Chapter 9: Hastie et al. (2001)

CN700 Additive Models and Trees Chapter 9: Hastie et al. (2001) CN700 Additive Mdels and Trees Chapter 9: Hastie et al. (2001) Madhusudana Shashanka Department f Cgnitive and Neural Systems Bstn University CN700 - Additive Mdels and Trees March 02, 2004 p.1/34 Overview

More information

Simple Linear Regression (single variable)

Simple Linear Regression (single variable) Simple Linear Regressin (single variable) Intrductin t Machine Learning Marek Petrik January 31, 2017 Sme f the figures in this presentatin are taken frm An Intrductin t Statistical Learning, with applicatins

More information

Sequential Allocation with Minimal Switching

Sequential Allocation with Minimal Switching In Cmputing Science and Statistics 28 (1996), pp. 567 572 Sequential Allcatin with Minimal Switching Quentin F. Stut 1 Janis Hardwick 1 EECS Dept., University f Michigan Statistics Dept., Purdue University

More information

ON-LINE PROCEDURE FOR TERMINATING AN ACCELERATED DEGRADATION TEST

ON-LINE PROCEDURE FOR TERMINATING AN ACCELERATED DEGRADATION TEST Statistica Sinica 8(1998), 207-220 ON-LINE PROCEDURE FOR TERMINATING AN ACCELERATED DEGRADATION TEST Hng-Fwu Yu and Sheng-Tsaing Tseng Natinal Taiwan University f Science and Technlgy and Natinal Tsing-Hua

More information

Math Foundations 20 Work Plan

Math Foundations 20 Work Plan Math Fundatins 20 Wrk Plan Units / Tpics 20.8 Demnstrate understanding f systems f linear inequalities in tw variables. Time Frame December 1-3 weeks 6-10 Majr Learning Indicatrs Identify situatins relevant

More information

SIZE BIAS IN LINE TRANSECT SAMPLING: A FIELD TEST. Mark C. Otto Statistics Research Division, Bureau of the Census Washington, D.C , U.S.A.

SIZE BIAS IN LINE TRANSECT SAMPLING: A FIELD TEST. Mark C. Otto Statistics Research Division, Bureau of the Census Washington, D.C , U.S.A. SIZE BIAS IN LINE TRANSECT SAMPLING: A FIELD TEST Mark C. Ott Statistics Research Divisin, Bureau f the Census Washingtn, D.C. 20233, U.S.A. and Kenneth H. Pllck Department f Statistics, Nrth Carlina State

More information

Comparing Several Means: ANOVA. Group Means and Grand Mean

Comparing Several Means: ANOVA. Group Means and Grand Mean STAT 511 ANOVA and Regressin 1 Cmparing Several Means: ANOVA Slide 1 Blue Lake snap beans were grwn in 12 pen-tp chambers which are subject t 4 treatments 3 each with O 3 and SO 2 present/absent. The ttal

More information

Homology groups of disks with holes

Homology groups of disks with holes Hmlgy grups f disks with hles THEOREM. Let p 1,, p k } be a sequence f distinct pints in the interir unit disk D n where n 2, and suppse that fr all j the sets E j Int D n are clsed, pairwise disjint subdisks.

More information

Computational modeling techniques

Computational modeling techniques Cmputatinal mdeling techniques Lecture 4: Mdel checing fr ODE mdels In Petre Department f IT, Åb Aademi http://www.users.ab.fi/ipetre/cmpmd/ Cntent Stichimetric matrix Calculating the mass cnservatin relatins

More information

CHAPTER 4 DIAGNOSTICS FOR INFLUENTIAL OBSERVATIONS

CHAPTER 4 DIAGNOSTICS FOR INFLUENTIAL OBSERVATIONS CHAPTER 4 DIAGNOSTICS FOR INFLUENTIAL OBSERVATIONS 1 Influential bservatins are bservatins whse presence in the data can have a distrting effect n the parameter estimates and pssibly the entire analysis,

More information

Kinetic Model Completeness

Kinetic Model Completeness 5.68J/10.652J Spring 2003 Lecture Ntes Tuesday April 15, 2003 Kinetic Mdel Cmpleteness We say a chemical kinetic mdel is cmplete fr a particular reactin cnditin when it cntains all the species and reactins

More information

Determining the Accuracy of Modal Parameter Estimation Methods

Determining the Accuracy of Modal Parameter Estimation Methods Determining the Accuracy f Mdal Parameter Estimatin Methds by Michael Lee Ph.D., P.E. & Mar Richardsn Ph.D. Structural Measurement Systems Milpitas, CA Abstract The mst cmmn type f mdal testing system

More information

Marginal Conceptual Predictive Statistic for Mixed Model Selection

Marginal Conceptual Predictive Statistic for Mixed Model Selection Open Jurnal f Statistics, 06, 6, 39-53 Published Online April 06 in Sci http://wwwscirprg/jurnal/js http://dxdirg/0436/js0660 Marginal Cnceptual Predictive Statistic fr Mixed Mdel Selectin Cheng Wenren,

More information

February 28, 2013 COMMENTS ON DIFFUSION, DIFFUSIVITY AND DERIVATION OF HYPERBOLIC EQUATIONS DESCRIBING THE DIFFUSION PHENOMENA

February 28, 2013 COMMENTS ON DIFFUSION, DIFFUSIVITY AND DERIVATION OF HYPERBOLIC EQUATIONS DESCRIBING THE DIFFUSION PHENOMENA February 28, 2013 COMMENTS ON DIFFUSION, DIFFUSIVITY AND DERIVATION OF HYPERBOLIC EQUATIONS DESCRIBING THE DIFFUSION PHENOMENA Mental Experiment regarding 1D randm walk Cnsider a cntainer f gas in thermal

More information

MATHEMATICS SYLLABUS SECONDARY 5th YEAR

MATHEMATICS SYLLABUS SECONDARY 5th YEAR Eurpean Schls Office f the Secretary-General Pedaggical Develpment Unit Ref. : 011-01-D-8-en- Orig. : EN MATHEMATICS SYLLABUS SECONDARY 5th YEAR 6 perid/week curse APPROVED BY THE JOINT TEACHING COMMITTEE

More information

IAML: Support Vector Machines

IAML: Support Vector Machines 1 / 22 IAML: Supprt Vectr Machines Charles Suttn and Victr Lavrenk Schl f Infrmatics Semester 1 2 / 22 Outline Separating hyperplane with maimum margin Nn-separable training data Epanding the input int

More information

I. Analytical Potential and Field of a Uniform Rod. V E d. The definition of electric potential difference is

I. Analytical Potential and Field of a Uniform Rod. V E d. The definition of electric potential difference is Length L>>a,b,c Phys 232 Lab 4 Ch 17 Electric Ptential Difference Materials: whitebards & pens, cmputers with VPythn, pwer supply & cables, multimeter, crkbard, thumbtacks, individual prbes and jined prbes,

More information

Statistical Learning. 2.1 What Is Statistical Learning?

Statistical Learning. 2.1 What Is Statistical Learning? 2 Statistical Learning 2.1 What Is Statistical Learning? In rder t mtivate ur study f statistical learning, we begin with a simple example. Suppse that we are statistical cnsultants hired by a client t

More information

[COLLEGE ALGEBRA EXAM I REVIEW TOPICS] ( u s e t h i s t o m a k e s u r e y o u a r e r e a d y )

[COLLEGE ALGEBRA EXAM I REVIEW TOPICS] ( u s e t h i s t o m a k e s u r e y o u a r e r e a d y ) (Abut the final) [COLLEGE ALGEBRA EXAM I REVIEW TOPICS] ( u s e t h i s t m a k e s u r e y u a r e r e a d y ) The department writes the final exam s I dn't really knw what's n it and I can't very well

More information

Slide04 (supplemental) Haykin Chapter 4 (both 2nd and 3rd ed): Multi-Layer Perceptrons

Slide04 (supplemental) Haykin Chapter 4 (both 2nd and 3rd ed): Multi-Layer Perceptrons Slide04 supplemental) Haykin Chapter 4 bth 2nd and 3rd ed): Multi-Layer Perceptrns CPSC 636-600 Instructr: Ynsuck Che Heuristic fr Making Backprp Perfrm Better 1. Sequential vs. batch update: fr large

More information

AP Statistics Notes Unit Two: The Normal Distributions

AP Statistics Notes Unit Two: The Normal Distributions AP Statistics Ntes Unit Tw: The Nrmal Distributins Syllabus Objectives: 1.5 The student will summarize distributins f data measuring the psitin using quartiles, percentiles, and standardized scres (z-scres).

More information

Lead/Lag Compensator Frequency Domain Properties and Design Methods

Lead/Lag Compensator Frequency Domain Properties and Design Methods Lectures 6 and 7 Lead/Lag Cmpensatr Frequency Dmain Prperties and Design Methds Definitin Cnsider the cmpensatr (ie cntrller Fr, it is called a lag cmpensatr s K Fr s, it is called a lead cmpensatr Ntatin

More information

Enhancing Performance of MLP/RBF Neural Classifiers via an Multivariate Data Distribution Scheme

Enhancing Performance of MLP/RBF Neural Classifiers via an Multivariate Data Distribution Scheme Enhancing Perfrmance f / Neural Classifiers via an Multivariate Data Distributin Scheme Halis Altun, Gökhan Gelen Nigde University, Electrical and Electrnics Engineering Department Nigde, Turkey haltun@nigde.edu.tr

More information

Module 3: Gaussian Process Parameter Estimation, Prediction Uncertainty, and Diagnostics

Module 3: Gaussian Process Parameter Estimation, Prediction Uncertainty, and Diagnostics Mdule 3: Gaussian Prcess Parameter Estimatin, Predictin Uncertainty, and Diagnstics Jerme Sacks and William J Welch Natinal Institute f Statistical Sciences and University f British Clumbia Adapted frm

More information

ENSC Discrete Time Systems. Project Outline. Semester

ENSC Discrete Time Systems. Project Outline. Semester ENSC 49 - iscrete Time Systems Prject Outline Semester 006-1. Objectives The gal f the prject is t design a channel fading simulatr. Upn successful cmpletin f the prject, yu will reinfrce yur understanding

More information

Emphases in Common Core Standards for Mathematical Content Kindergarten High School

Emphases in Common Core Standards for Mathematical Content Kindergarten High School Emphases in Cmmn Cre Standards fr Mathematical Cntent Kindergarten High Schl Cntent Emphases by Cluster March 12, 2012 Describes cntent emphases in the standards at the cluster level fr each grade. These

More information

Discussion on Regularized Regression for Categorical Data (Tutz and Gertheiss)

Discussion on Regularized Regression for Categorical Data (Tutz and Gertheiss) Discussin n Regularized Regressin fr Categrical Data (Tutz and Gertheiss) Peter Bühlmann, Ruben Dezeure Seminar fr Statistics, Department f Mathematics, ETH Zürich, Switzerland Address fr crrespndence:

More information

INSTRUMENTAL VARIABLES

INSTRUMENTAL VARIABLES INSTRUMENTAL VARIABLES Technical Track Sessin IV Sergi Urzua University f Maryland Instrumental Variables and IE Tw main uses f IV in impact evaluatin: 1. Crrect fr difference between assignment f treatment

More information

We can see from the graph above that the intersection is, i.e., [ ).

We can see from the graph above that the intersection is, i.e., [ ). MTH 111 Cllege Algebra Lecture Ntes July 2, 2014 Functin Arithmetic: With nt t much difficulty, we ntice that inputs f functins are numbers, and utputs f functins are numbers. S whatever we can d with

More information

Name: Block: Date: Science 10: The Great Geyser Experiment A controlled experiment

Name: Block: Date: Science 10: The Great Geyser Experiment A controlled experiment Science 10: The Great Geyser Experiment A cntrlled experiment Yu will prduce a GEYSER by drpping Ments int a bttle f diet pp Sme questins t think abut are: What are yu ging t test? What are yu ging t measure?

More information

SAMPLING DYNAMICAL SYSTEMS

SAMPLING DYNAMICAL SYSTEMS SAMPLING DYNAMICAL SYSTEMS Melvin J. Hinich Applied Research Labratries The University f Texas at Austin Austin, TX 78713-8029, USA (512) 835-3278 (Vice) 835-3259 (Fax) hinich@mail.la.utexas.edu ABSTRACT

More information

BOUNDED UNCERTAINTY AND CLIMATE CHANGE ECONOMICS. Christopher Costello, Andrew Solow, Michael Neubert, and Stephen Polasky

BOUNDED UNCERTAINTY AND CLIMATE CHANGE ECONOMICS. Christopher Costello, Andrew Solow, Michael Neubert, and Stephen Polasky BOUNDED UNCERTAINTY AND CLIMATE CHANGE ECONOMICS Christpher Cstell, Andrew Slw, Michael Neubert, and Stephen Plasky Intrductin The central questin in the ecnmic analysis f climate change plicy cncerns

More information

The general linear model and Statistical Parametric Mapping I: Introduction to the GLM

The general linear model and Statistical Parametric Mapping I: Introduction to the GLM The general linear mdel and Statistical Parametric Mapping I: Intrductin t the GLM Alexa Mrcm and Stefan Kiebel, Rik Hensn, Andrew Hlmes & J-B J Pline Overview Intrductin Essential cncepts Mdelling Design

More information

1 The limitations of Hartree Fock approximation

1 The limitations of Hartree Fock approximation Chapter: Pst-Hartree Fck Methds - I The limitatins f Hartree Fck apprximatin The n electrn single determinant Hartree Fck wave functin is the variatinal best amng all pssible n electrn single determinants

More information

22.54 Neutron Interactions and Applications (Spring 2004) Chapter 11 (3/11/04) Neutron Diffusion

22.54 Neutron Interactions and Applications (Spring 2004) Chapter 11 (3/11/04) Neutron Diffusion .54 Neutrn Interactins and Applicatins (Spring 004) Chapter (3//04) Neutrn Diffusin References -- J. R. Lamarsh, Intrductin t Nuclear Reactr Thery (Addisn-Wesley, Reading, 966) T study neutrn diffusin

More information

Pure adaptive search for finite global optimization*

Pure adaptive search for finite global optimization* Mathematical Prgramming 69 (1995) 443-448 Pure adaptive search fr finite glbal ptimizatin* Z.B. Zabinskya.*, G.R. Wd b, M.A. Steel c, W.P. Baritmpa c a Industrial Engineering Prgram, FU-20. University

More information

Differentiation Applications 1: Related Rates

Differentiation Applications 1: Related Rates Differentiatin Applicatins 1: Related Rates 151 Differentiatin Applicatins 1: Related Rates Mdel 1: Sliding Ladder 10 ladder y 10 ladder 10 ladder A 10 ft ladder is leaning against a wall when the bttm

More information

Smoothing, penalized least squares and splines

Smoothing, penalized least squares and splines Smthing, penalized least squares and splines Duglas Nychka, www.image.ucar.edu/~nychka Lcally weighted averages Penalized least squares smthers Prperties f smthers Splines and Reprducing Kernels The interplatin

More information

UNIV1"'RSITY OF NORTH CAROLINA Department of Statistics Chapel Hill, N. C. CUMULATIVE SUM CONTROL CHARTS FOR THE FOLDED NORMAL DISTRIBUTION

UNIV1'RSITY OF NORTH CAROLINA Department of Statistics Chapel Hill, N. C. CUMULATIVE SUM CONTROL CHARTS FOR THE FOLDED NORMAL DISTRIBUTION UNIV1"'RSITY OF NORTH CAROLINA Department f Statistics Chapel Hill, N. C. CUMULATIVE SUM CONTROL CHARTS FOR THE FOLDED NORMAL DISTRIBUTION by N. L. Jlmsn December 1962 Grant N. AFOSR -62..148 Methds f

More information

Margin Distribution and Learning Algorithms

Margin Distribution and Learning Algorithms ICML 03 Margin Distributin and Learning Algrithms Ashutsh Garg IBM Almaden Research Center, San Jse, CA 9513 USA Dan Rth Department f Cmputer Science, University f Illinis, Urbana, IL 61801 USA ASHUTOSH@US.IBM.COM

More information

SURVIVAL ANALYSIS WITH SUPPORT VECTOR MACHINES

SURVIVAL ANALYSIS WITH SUPPORT VECTOR MACHINES 1 SURVIVAL ANALYSIS WITH SUPPORT VECTOR MACHINES Wlfgang HÄRDLE Ruslan MORO Center fr Applied Statistics and Ecnmics (CASE), Humbldt-Universität zu Berlin Mtivatin 2 Applicatins in Medicine estimatin f

More information

B. Definition of an exponential

B. Definition of an exponential Expnents and Lgarithms Chapter IV - Expnents and Lgarithms A. Intrductin Starting with additin and defining the ntatins fr subtractin, multiplicatin and divisin, we discvered negative numbers and fractins.

More information

Building to Transformations on Coordinate Axis Grade 5: Geometry Graph points on the coordinate plane to solve real-world and mathematical problems.

Building to Transformations on Coordinate Axis Grade 5: Geometry Graph points on the coordinate plane to solve real-world and mathematical problems. Building t Transfrmatins n Crdinate Axis Grade 5: Gemetry Graph pints n the crdinate plane t slve real-wrld and mathematical prblems. 5.G.1. Use a pair f perpendicular number lines, called axes, t define

More information

Revision: August 19, E Main Suite D Pullman, WA (509) Voice and Fax

Revision: August 19, E Main Suite D Pullman, WA (509) Voice and Fax .7.4: Direct frequency dmain circuit analysis Revisin: August 9, 00 5 E Main Suite D Pullman, WA 9963 (509) 334 6306 ice and Fax Overview n chapter.7., we determined the steadystate respnse f electrical

More information

How do scientists measure trees? What is DBH?

How do scientists measure trees? What is DBH? Hw d scientists measure trees? What is DBH? Purpse Students develp an understanding f tree size and hw scientists measure trees. Students bserve and measure tree ckies and explre the relatinship between

More information

the results to larger systems due to prop'erties of the projection algorithm. First, the number of hidden nodes must

the results to larger systems due to prop'erties of the projection algorithm. First, the number of hidden nodes must M.E. Aggune, M.J. Dambrg, M.A. El-Sharkawi, R.J. Marks II and L.E. Atlas, "Dynamic and static security assessment f pwer systems using artificial neural netwrks", Prceedings f the NSF Wrkshp n Applicatins

More information

We say that y is a linear function of x if. Chapter 13: The Correlation Coefficient and the Regression Line

We say that y is a linear function of x if. Chapter 13: The Correlation Coefficient and the Regression Line Chapter 13: The Crrelatin Cefficient and the Regressin Line We begin with a sme useful facts abut straight lines. Recall the x, y crdinate system, as pictured belw. 3 2 1 y = 2.5 y = 0.5x 3 2 1 1 2 3 1

More information

Inference in the Multiple-Regression

Inference in the Multiple-Regression Sectin 5 Mdel Inference in the Multiple-Regressin Kinds f hypthesis tests in a multiple regressin There are several distinct kinds f hypthesis tests we can run in a multiple regressin. Suppse that amng

More information

STATS216v Introduction to Statistical Learning Stanford University, Summer Practice Final (Solutions) Duration: 3 hours

STATS216v Introduction to Statistical Learning Stanford University, Summer Practice Final (Solutions) Duration: 3 hours STATS216v Intrductin t Statistical Learning Stanfrd University, Summer 2016 Practice Final (Slutins) Duratin: 3 hurs Instructins: (This is a practice final and will nt be graded.) Remember the university

More information

(1.1) V which contains charges. If a charge density ρ, is defined as the limit of the ratio of the charge contained. 0, and if a force density f

(1.1) V which contains charges. If a charge density ρ, is defined as the limit of the ratio of the charge contained. 0, and if a force density f 1.0 Review f Electrmagnetic Field Thery Selected aspects f electrmagnetic thery are reviewed in this sectin, with emphasis n cncepts which are useful in understanding magnet design. Detailed, rigrus treatments

More information

Preparation work for A2 Mathematics [2017]

Preparation work for A2 Mathematics [2017] Preparatin wrk fr A2 Mathematics [2017] The wrk studied in Y12 after the return frm study leave is frm the Cre 3 mdule f the A2 Mathematics curse. This wrk will nly be reviewed during Year 13, it will

More information