Basic Nonparametric Estimation Spring 2002

Size: px
Start display at page:

Download "Basic Nonparametric Estimation Spring 2002"

Transcription

1 Basic Nonparametric Estimation Spring 2002 Te following topics are covered today: Basic Nonparametric Regression. Tere are four books tat you can find reference: Silverman986, Wand and Jones995, Hardle990, Hardle994C 38 in te Handbook of Econometrics vol 4. Key Words: Density estimate, conditional expectation, kernel smooting, k-nnk nearest neigborood, local polynomial, seriessieve, spline, curse of dimensionality, optimal rate of convergence, bias-variance trade-off, undersmooting, oversmooting, pointwise asymptotic distribution, pointwise confidence interval, uniform confidence interval, automatic selection of bandwidt, uniform rate of convergence. Basic View: Tere can be many meanings to nonparametrics. One meaning is optimization over a set of function. For example, given te sample of observations x,..., x n, you are asked to find a distribution function under wic te joint probability of x,..., x n is maximized. You can convince yourself tat tis is te empirical distribution: Essentially, you are looking for probability masses on x,..., x n, say p,..., p n, suc tat p p n = and p p n is maximized. Convince yourself tat te solution is p =... = p n =. Tis is n also called nonparametric maximum likeliood. But tis is not wat we will discuss ere. Te meaning of nonparametric for now is density estimate and estimation of conditional expectations. Density estimate: motivation For tis part you sould read te first several pages in Hardle994. One motivation is tat you first use te istogram to estimate te density: #ofx i in x, x + = 2 n 2 n t= x x i x + = i= x 2 xi Now x is te uniform density over,, tis is called te uniform kernel. Of 2 course you can replace by anoter density function K, not neccessary over,, and plug in into were te uniform density appears, to get ˆf x = n t= K x x i. Anoter motivation, probably easier to understand: you estimate te distribution function F x by ˆF x = n n t=] x i x but you can t differentiate it to get te density. So you smoot out ˆF x by replacing x i x by G x i x were G is any smoot distribution function so tat G =, G = 0, and 0. In practice, you just take to be some small but fixed number, like 0.. So you ave ˆF x = n n t= G x i x, differentiate tis to get ˆf x = n t= K x i x for K = G te density corresponding to G. Higer dimension: if x R d a d-dimension vector, ten just take G to be a d-dimensional distributional and do te same ting, ˆF x = n n t= G x i x. Now wen you differentiate it don t forget tat you ave to do tat to eac of te x,..., x d, so you will pull out d of to get: ˆf x = d xi x K t=

2 were now K is te multivariate density corresponding to te multivariate distribution function G. Conditional expectation: motivation You want to estimate E y x or more generally E g y x for some function g. You can also estimate tings like te conditional median of y give x, say med y x, or any oter conditional quantiles. Local Weigting: Te wole ting about nonparametric estimate of E y x is just use observations for wic x i is close to x. Te obvious to do is to take a neigborood around x and average over tose y i for wic x i is wit tis neigborood. Te size of te neigborood sould srink to 0 but not too fast. More generally you take weigted average over all y i but you give more weigts to tose y i for wic te x i tat is close to x, and you give less weigt to tose y i for wic te x i is far away from x. See Stone977. For weigts W n x, x i suc tat i= W n x, x i = tis is wat weigt means; 2 W n x, x i 0 if x i x; 3 max i n W n x, x i 0 as n, you estimate E y x by i= W n x, x i Y i. Similarly, you estimate E y 2 x by i= W n x, x i Yi 2, and F y x by i= W n x, x i Y i y. Anytings you do parametrically, if you do tat only for x i close to x, ten you become nonparametric. You can run least square, quantile regression, and even maximum likeliood over smallsrinking neigborood of x, and call tat nonparametric st st... Depending on ow you define te weigt sequence W n x, x i, te local weigting type nonparametric estimators are classified into many many different categories. Te most important ones are kernel smooting, k-nearest neigborood, local polynomials. Asymptotically tey are all equivalent to eac oter. Kernel smooting is te easiest to use so tis is te only ting we will deal wit in details. In additions to tese local nonparametric estimate, tere are also global nonparametric estimates including seriessieve estimator and splines. Seriessieve estimation as been paid increasingly attention recently. But few people would boter wit splines. Te reason again is tat seriessieve is extremely easy to do, you just need to run least square! And just remember to say tat te number of regressors included as n, altoug in practice you just include some finite number of terms. But te focus today is kernel. Kernel smooting: If you use density weigting for te weigts W n x, x i, ten you get te kernel estimator of E y x. If x i is one-dimensional, let W n x, x i = i= i= K x x i i= K x x i tis very definition i= W n x, x i =. Ten te kernel estimate is defined by: W n x, x i Y i = K x x i i= K i= x x i Y i = K x x i Yi i= K x x i For multidimension x i R d, use te multidimension density estimate W n x, x i = and te kernel estimator of E y x is given by: W n x, x i Y i = i= i= K x x i d d i= K x x i Y i = 2 Yi d i= K x x i d i= K x x i. By K x x d i K x x d i i=,

3 Anoter view of kernel estimator: You can tink of E y x = E y x f x f x = γ x f x were γ x E y x f x = yf y, x dy. So you may estimate γ x and f x seperately. Of course ˆf x = n d i= K x x i. For ˆγ x, plug in te kernel density estimate for x, y: ˆf x, y = n d+ i= K x x i K yi y into yf y, x dy, and use a cange of variable u = y i y /: y ˆf y, x dy = d = d x xi K x xi K i= i= y K yi y dy y i + u K u du = d x xi K y i Yet anoter view for γ x: you can tink of y ˆf y, x dy as ydp, were P is te measure over y defined by P y i y, x i = x. Note tat P y i y, x i = x = d P y dx i y, x i x, you can estimate P y i y, x i x by n n i= y i y G x i x /. Differentiating tis, you estimate d P y dx i y, x i x by n d i= y i y K x i x /. Plug in tis estimate of P into ydp : yd ˆP = = d yd d y i y K x i x / i= K x i x / i= yd y i y = d i= K x i x / y i Note: Now note tat to study convergence rate, asymp distribution, etc, we only need to worry about ˆγ x, and we may ignore discussing ˆf x. Te reason is tat ˆf x is just a special case of ˆγ x, in wic you take y i identically. Convenient forms of kerneldensity function include tose on pp2303 Hardle994 C8 vol Handbook: Uniform kernel x ; 2 Triangular kernel: u u ; and 2 quartic, epanecniknov, gaussian, etc. Read pp2303, and try to use gauss or matlab to grap out wat eac of tese look like. To get consistency, you want 0 as n 0 so tat you don t ave a bias, you also want d so tat you ave enoug observations. More on tis below. Estimating Derivatives: It is also easy to estimate derivatives of γ xincluding f x. As long as your kernel is smoot differentiable, you could just simply differentiate your ˆγ x and f x. To estimate te kt derivative of γ x, denoted by γ k x, use ˆf k x = k+d i= 3 i= K k xi x y i

4 Don t forget tat eac time you differentiate wrt x in K x i x, you will pull out one, so altogeter you pull out k of tem. Also remember tat wen x is d-dimension, te derivative can be taken wit respect to different combinations out of x,..., x d, as long as te degree of differentiation wit respect to x,..., x d sum to k. Sometimes people write f k and similarly K k using very long notation. For example, f k is written as: for positive integers λ,..., λ d, let λ λ d = k. Ten write f k x as f x. k x λ... xλ d d k-nnnearest neigborood and local polynomials: Tese are te oter two major weigting scemes for W ni x. Bot of tese are asymptotically equivalent to kernel weigting. k-nn: Insteading of using observations wose x i are wit far away from x, you may also just use your k closest neigbors, assuming you are standing at te point x. Look at te example in pp3 Hardle990 for te precise definition. You may weigt your k nearest neigbors equally, or you may give even iger weigt to te ones wo just live next to you and give decreasing weigts to te ones increasing furter away from you. You can of course use any kernel density weigt K. In tis case you may take te bandwidt to be te distance between x and te kt nearest neigbor x i. Giving equal weigts to eac of te k neigbors is te same as saying using an uniform kernel density weigt. If you do tis ten tere is an easy relation between and k: k n 2 d f x. For te case were d =, ceck to see tat tis relation makes sense to you by verfitying te equivalence between kernel and k-nn in Table 3.2. pp46 Hardle990. For local polynomial, read pp30-32 Hardle990. You must know ow to run a polynomial regression using all observations, just regress y on, x, x 2, all cross products, cross products of all powers..., up to degree k. If you run tis kt degree polynomial regression only over a neigborood for wose x i wit distance from x: x i x, ten you get te local polynomial estimate. Te degreee k of local polynomial corresponding to te order of te kernel, wic we will talk about below. Series and Splines: Tere is also anoter ting called sieve estimator, wic is very similar to, and you can tink tat it is just same as, series estimator. Te only difference between series and local polynomials is tat you run te polynomials using all observations, instead of only a srinking neigborood x, x +. However, now instead of fixing k, te degree of polynomial, you promise to let k. But tis is noting more tan a promise, you don t actually do it. Just take a number k, not too small. Instead of using polynomials, wic tends to be igly correlated, you may also use family of ortogonal series of functions, like trigonometric function, legendre polynomials, laguerre polynomials. If you are interested you can find more information about tese function in a recent book by Judd998. Splines is based on find a twice differentiable function g x tat minimizes i= y i g x i 2 + λ g x 2 dx, for some λ > 0. Te reason to include te second term is to penalize te rougness of te estimate g, oterwise te minization problem is ill-defined and you will simply take ĝ x i = y i, no good. Tis will give you a cubic polynomial wit continuous second derivatives. Of course instead of penalizing by λ g x 2 dx, you migt tink of all oter kinds of penalizations. But tis as not been muc used oter tan simple case wit one dimension x. If you are interested read Hardle990 sec 3.4 pp56. 4

5 Now back to te local nonparametric estimators. Curse of dimensionality: For a given bandwidtwindow size, te iger dimension x, te less data you ave in a neigborood wit bandwidt. Say all your x is on te unit interval 0,, you only look at observations witin a small bandwidt, te total number of observations is n. If d =, ten te effective number of observations you ave in tat neigborood is. If d = 2, te effective number of observations you ave is 2, so on so fort. So for general d, te number of observations you ave in a d dimension cube wit lengt is just d. Since is small, goes to 0, of course you get less and less observations as d get larger. Anoter view of tis is tat, fixing te number of total observations, te average distance between any pair of observations gets larger and larger as d increasesyou find tat your data becomes sparser and sparser in iger dimensions. Optimal rate of convergence for nonparametric estimates: Suppose te true function γ x is pt degree differentiable, all pt derivative bounded uniformly over x. Ten te optimal bandwidt opt you sould use is n, and te best rate at wic your estimate ˆγ x can approac γ x is O p n p. Now we will explain wy. Reading is in sec 4 of Hardle990. You may consult te original Stone papers if you are interested. Te problem ere is te bias and variance tradeoff. Basically, by using a bandwidt of size you are doing two tings: you are using observations x i tat can be at most away from x to approximate te true function γ x at point x; 2 you are using all d observations in tis small cube of lengt. Step gives you a bias in estimating γ x, and step 2 gives you te variance in te estimate. Te smaller te, te smaller te bias, but te less observations you ave, tus te large te variance. Te larger te, te larger te bias, but ten you ave more observations wic reduces te variance in your estimate. To get consistency, you want to get rid of bot te bias and te variance. To get rid of bias, simply take 0, as n 0. To get rid of variance, you want to ave infinite number of observations as n, but te number of observations you ave is d, so you also want d. If you ave bot 0 and d, ten your estimate is consistent. However, you want more tan consistency. You want ˆγ x to converge as fast as possible, to do tat you need to balance te trade off between bias and variance. To do tis, we need to find more precise representations of te bias and te variance. Te bias is of order p : Because you know your function γ x ave p bounded derivatives, so you could do a taylor expansion of te approximation error γ x i g x around x up to degree p. Now eac term is a polynomial in x i x and remember tat x i x. Just assume tat you asomeow knows te coefficients in tose p expansion terms, so tat you don t ave to worry about tese terms. But te last term, te pt order term, an order of x i x p p cannot be furter expanded anymore, so you ave to live up wit it. Te order of tis error is p. Tese are different ways to get rid of te first p terms in te taylor expansion. Higer order kernel is one way, wic we will see below. Use p t order local polynomial is anoter waywic is just exactly doing a taylor expansion of order p but we won t get into tis. Te variation is of order O p. Just tink of te simplest problem of estimating a d 5

6 mean using a sample average. Wen you ave n observations, using a central limit teorem, you know tat d n x µ N 0,? = O p. So tat x µ = O p n. Now wit a window size te number of observations you ave is d, by analogy to te simple case of x, te precision of your estimate, of course, sould be O p. d So tat you total error, bias plus estimation error, is of te order of p +, find a to d minize tis, you can easily see tat opt = O n. A more formal way to tink of te total error is MSEmean square error = bias 2 + variance. Now bias 2 = 2p, and variance =. We will see tese in detail for kernel regressions. Ten te pointwise optimal rate of convergence d is given by: O p opt = O = O n p d It sould ten be clear tat it is not possible to ave n convergence for nonparametric p estimates since <. Sometimes 2 n/4 rate of convergence is needed for getting rid of te p second order terms for semiparametric estimators. You will need p > d/2 to get > /4. So te more regressors you ave, te more smoot your regression function as to be in order to get te same rate of convergence. Optimal rate for derivative estimates: Wat is te optimal bandwidt for estimating γ k x and wat is te best convergence rate. It turns out te optimal bandwidt is of te same order as tat of estimating γ x itself. Tis can be seen again from te same bias and variance tradeoff. bias: Now if γ x as p bounded derivatives, ten γ k x as p k bounded derivatives. So by analogy, te bias is of te order of p k. variance: Again you ave d observations effectively. However, eac differentiation amplify te error by an order of so after taking k derivative you amplify te error by, see k equation. So te estimation error is of te order of k. Te total error is of order d p k + k. Find a to minimize tis again you see tat d opt = n. Note tat opt does not depend on k at all. Ten te best convergence rate is: n p k = k = d n Te order of bias and variance will be developed in detail for kernel estimates below. Higer order kernels Te kernel weigts sould integrate to, as required for a density, so K u du =. If you use a density K tat is symmetric around 0, ten uk u du = 0. But if you insist tat te kernel K must be a density, ten K u 0, u, and so u 2 K u du > 0. A iger order kernel is tose K tat as some similarity to a symmetric density but is not a density. Precisely, for tese K, K u udu = 0, but may be negative for some u. Tis allows you to find suc K so tat K u u 2 du = 0. To see graps of example, see 6 p k

7 e.g figure 3. and 3.2 in pp60 and pp6 in Hardle990. Precisely, a kernel of order r is defined as tose K for wic: K u du = K u u q du = 0, q =,..., r u r K u du < for examples of tese kernel can be found in table 4.5. pp35 Hardle990. Bias of kernel estimates: use te cange of variable u = x i x, and bear in mind tat te integration notation is over te d-dimensional x i, and finally do a taylor expansion of γ x + u, Eˆγ x γ x =E x xi d K Y i γ x = E x d K xi Y i γ x i= x =E d K xi x m x i f x i dx i γ x = E d K xi γ x i dx i γ x r = K u γ x + u du γ x = j γj u j k u du + r γ r x u r K u du j! r! j= If your kernel is of te order r, ten eac of te integrals u j k u du = 0 in te first summation. So you are left wit only te last term, so tat bias = γ. Note tat you can do tis only if your true underlying function γ x as p r bounded derivatives. So if your function γ x as pt bounded derivatives and your kernel is of order r, ten te bias = minp,r. Optimally, if you know your γ x is pt degree smoot, ten you sould use a pt degree kernel to reduce bias. But you never know wat p-smootness your function is. So wy not always use a very ig order kernel? Tat is because a ig order kernel will neccessary make te variance big, as seen below. In te basic case wen you just use a symmetric density for K, it is of order r = 2 so tat te bias is of order 2. Variance of kernel estimates: Tis one is easier tan te bias since iger order kernel does not play any role ere, let g x i = E yi 2 x i f x i : x xi V ar d K Y i i= = x xi n 2 2d V ar K Y i = x 2d E K 2 xi Yi 2 x 2 xi 2d EK Y i i= = x xi d d K2 E y 2 i x i f xi dx i E x 2 n d K xi Y i = x xi d d K2 g x i dx i + O = n d K 2 u g x + u du + O n = d K 2 u g x du + d K 2 u g x udu + O n = d K 2 u g x du + O d + O = O n d 7

8 Bias and Variance of derivatives of kernel estimates: You could verify te formulas for bias and variance for γ k x by carrying troug te same exercise of taking expectation, cange of variable in integration. An additional tools tat are needed ere is integration by parts, assuming K u is bounded and all te first kt derivatives dimins at te boundary, you can verify tat: K k u γ u du = k K u γ k u du. Exercise: Use integration by parts to obtain te bias formula for kernel estimates of γ k x, and convince yourself tat iger order kernel plays te same role ere. Te variance formula can be calculated as before witout aving to worry about integration by parts. Asymptotic Distribution, Confidence band To sow ow precise your nonparametric estimate is at eac point x, it is neccessary to derive its asymptotic distribution and obtain te confidence interval at eac point x. Section 4 in Hardle990 contains detailed discussion. Te derivation of te asymptotic distribution depends on wat bandwidt you use, if you use opt, were means te same order as, ten te asymptotic distribution will depend on bot te bias and te variance. On te oter and, if you use << opt were << means smaller by an order of magnitude, i.e., opt 0, ten tere is no bias in te asymptotic distribution but te convergence rate is not te fastest. To look at tis in more detail, consider te simple example for d =, r = 2, for one-dimensional x and for a standard density kernel wit only order 2, ten opt = n = n 5. To compute te asymptotic distribution of opt ˆm x m x = 2 opt ˆm x m x, for ˆm x = ˆγx ˆfx, first linear tis in terms of ˆγ x and ˆf x by simply taking first order taylor expansion: ˆm x m x γ x ˆγ x γ x ˆf x f x f x f x 2 As seen above: Eˆγ x γ x = 2 2 γ x u 2 K u du. Note tat for γ x = m x f x, γ x = m x f x + 2f x m x + m x f x. So for te density were m x, E ˆf x f x = 2 2 f x. Terefore E 2 opt ˆm x m x = γ x 2 f = 2 m f f x = 2 f 2m x f x + m x f x f x m f + 2m f + mf m f f x u 2 K u du. To compute te variance, follows from te variance calculation for kernel estimation to get: V ar ˆγ x γ x g x K 2 u du, for g x = E y 2 x f x. For density estimate were y, g x = f x, so tat V ar ˆf x f x g x K 2 u du. It is also neccessary to calculate te covariance between ˆγ x and ˆf x, you may use te type of cange-of-variable calculation to get: Cov ˆγ x γ x, ˆf x f x γ x K 2 u du. u 2 K u du 8

9 Ten use te delta metod to obtain te variance of V ar ˆm x m x : V ar ˆm x m x = V ar = K 2 u du f ˆγ x m f ˆf f 2 E y 2 x f x 2 m2 m x γ x + f f f = E y 2 x m x 2 K 2 u du = f x f x σ2 x K 2 u du To summarize: ˆm x m x d m x f x + 2m x f x N 2f x u 2 K u du, f x σ2 x K 2 u du If instead of using opt = n /5, you use a undersmoot bandwidt << n /5, say = n /6, ten te bias term will disappear and you don t ave to worry about it anymore. In tat case te asymptotic distribution simplies to: d ˆm x m x N 0, f x σ2 x K 2 u du Terefore, if you don t want to boter wit te complicated bias term in drawing your confidence interval, ten you sould use a undersmoot bandwidt, like = n /4. However, if you insist on using te optimal opt and want to draw te confidence interval around your estimated ˆm x, ten you will need to be able to estimate te bias term consistently. However, te bias term involve second derivatives m x, or for tat matter γ x. And γ x can NOT be estimated consistently using opt. Instead, you will need to use a seperate and oversmooted bandwidt, say g = n /6 to estimate γ x consistently. Te reason is simply tat ˆγ x as a bigger variance tan ˆγ x, by an order of 2. To reduce tis larger bias you need more data and so you need a larger oversmooted bandwidt g. In fact you can sow tat te variance of ˆγ x is O /5, so if you use opt = n /5, te variance won t go to 0. But if you use g = n /6, ten it will. Tis g is called pilot bandwidt in Hardle990 s terminology. Uniform confidence band: Te above asymptotic distribution is for deriving te Pointwise confidence band. Te uniform band is te two curves for wic te probability tat tey encompass te ENTIRE true m x function is α%. Constructing uniform band is difficult. You migt consult Hardle990 sec 4.3 if you are interested. But I suggest against doing tat in first reading. You need to know owever, te difference between pointwise confidence band and uniform band. Automatic bandwidt selection: Two ways: cross validation and penalizing functions, read Hardle990 pp Cross Validation: You want a good fit of your estimate, try to minimize i= ˆm x i m x i 2, but you don t know te true m x i, so you estimate it using y i : min i= ˆm x i y i 2. But 9

10 ten you are in trouble, since as 0, ˆm x i = y i, so you ave perfect fit 0, same problem tat motivates te spline estimator. Anoter way to tink about tis, write ˆm x i y i 2 = i= ˆm x i m x i ɛ i 2 = i= ˆm x i m x i 2 + ɛ 2 i 2 i= ˆm x i m x i ɛ i Te first term is wat you want. Te second term is not related to and can be ignored. Te tird term causes trouble, its expectation: E i= xi x j K j= ɛ j ɛ i = i= K 0 σ 2 = σ2 K 0 2 Tis is not zero and in fact not negligiblesame order as first term. Te problem is due to te correlation between ɛ i and itself. Of course if ɛ i is not used to estimate m x i ten tis term wouldn t exist to cause problem, terefore leading to te leave-one-out estimate ˆm i x i = n j=,j i K x j x i i= yi, and define te cross-validation function tat is to be minimized over as: min CV = i= m i x i m x i 2. According to Hardle990 sec 5. tis will automatically give you te optimal bandwidt, no matter ow smoot your m x is. Of course tis statement needs to be constrained by te order of te kernel function you use. Terefore te point of cross-validation is to get rid of te nonzero expectation due to correlation of ɛ i wit itself, in 2, anoter way to get rid of tis term is just to subtract it, if you know σ 2. But you don t know σ 2, but you can still estimate tis term consistently by K 0 n i= y i ˆm x i 2. Tis leads to te penalizing function metods, read pp54 Hardle990, to minimizer over : G = ˆm x i y i 2 + 2K 0 n i= y i ˆm x i 2 i= G as te same property as CV. Bias reduction by Jacknifing: skip te rest of tis note if you don t find it interesting. Tese are Extra material Anoter way to tink about bias reduction is by Jacknifing. Read Hardle990 sec 4.6in particular page 43 for details. It is essentially equivalent to ig order kernel. It doesn t make any difference if you are just running a simple kernel regression. So if you are just running a simple kernel regression don t boter wit Jacknifing. However, if you ave some objective function tat are only convex wit positive K, say, if you want to run a nonparametric quantile regression and want to reduce bias, ten operationally te Jacknife metod is very useful in preserving te convexity of te objective function. See for example, te recent paper Honore and Powell998 for illustration. Uniform rate of convergence: In addition to te pointwise rate. It is sometimes useful to obtain optimal bandwidt and optimal uniform convergence rate, i.e., for sup x X ˆγ x γ x. Te necceesary steps are again to consider te bias-variance tradeoff. Te bias for sup x X ˆγ x γ x for a rt order kernel is te same before p, noting tat te bias calculation is completely uniform in x. Te variance calculation, on te oter and, need to 0

11 /2 be modified. It turns out te variance is of te order O d p Eˆγ x = O p d, i.e. sup x X ˆγ x /2. Te mysterious comes from considering te variance at an increasing number of points simultaneous and by te application of te Bernstein inequality, a form of te exponential inequality: Berstein inequality: Tis is one of te many exponential inequality used to bound te tail probability of sums of mean 0 random variables. To see a clean and brief discussion of it you may read pp 93 of Pollard984. It basically says tat for z,..., z n independent of eac oter, eac z i bounded by B, V = V ar z V ar z n, S n = z z n, η P S n > η 2 exp 2. Tis says tat S n looks rougly like normal distribution 2 V + 3 Bη wit variance V in te tails. Te term Mη turns out to be not important and may in fact 3 be ignored. /2, Variance for uniform rate: To sow tat sup x X ˆγ x Eˆγ x = O p d te first ting is to partition te compact set X into side lengt n L squares, for some big number L. If X is compact, say just te unit cube, ten you got about n dl suc squares wit side lengt n L. Let x j, j =,..., n dl be te center of eac of te squares. As long as L is big enoug, te difference between γ x Eˆγ x and γ x 2 Eˆγ x 2 for x and x 2 in te same square is small enoug and may be ignored, because now te difference is of order n L d /2. << d Ten rougly speaking, supx X ˆγ x Eˆγ x sup xj,j=,...,n ˆγ x j Eˆγ x dl j. Now for eac x j, ˆγ x j Eˆγ x j is sums of mean 0 random variables, te sum of variance V, as ave been calculated above, is about D, for some constant D. Assuming bot Y d i and K are bounded, ten B = E for some E > 0. remember also tat saying tat d d /2 sup ˆγ x Eˆγ x = O p x X means lim d /2 P M sup ˆγ x Eˆγ x > M 0. n x X Now time to use te Bernstein inequality: d /2 d /2 P sup ˆγ x Eˆγ x > M < P sup ˆγ x j Eˆγ x j > M x X x j,j=,...,n dl n dl 2 exp M 2 d n dl 2 exp M2D 2 = 2 exp dl M M,n 0. 2D D + E d 3 M d d Ten minimize te sum of bias + error = p + O p d bandwidt and te best convergence rate. /2 you will find te optimal

The Priestley-Chao Estimator

The Priestley-Chao Estimator Te Priestley-Cao Estimator In tis section we will consider te Pristley-Cao estimator of te unknown regression function. It is assumed tat we ave a sample of observations (Y i, x i ), i = 1,..., n wic are

More information

Chapter 1. Density Estimation

Chapter 1. Density Estimation Capter 1 Density Estimation Let X 1, X,..., X n be observations from a density f X x. Te aim is to use only tis data to obtain an estimate ˆf X x of f X x. Properties of f f X x x, Parametric metods f

More information

Material for Difference Quotient

Material for Difference Quotient Material for Difference Quotient Prepared by Stepanie Quintal, graduate student and Marvin Stick, professor Dept. of Matematical Sciences, UMass Lowell Summer 05 Preface Te following difference quotient

More information

lecture 26: Richardson extrapolation

lecture 26: Richardson extrapolation 43 lecture 26: Ricardson extrapolation 35 Ricardson extrapolation, Romberg integration Trougout numerical analysis, one encounters procedures tat apply some simple approximation (eg, linear interpolation)

More information

Numerical Differentiation

Numerical Differentiation Numerical Differentiation Finite Difference Formulas for te first derivative (Using Taylor Expansion tecnique) (section 8.3.) Suppose tat f() = g() is a function of te variable, and tat as 0 te function

More information

REVIEW LAB ANSWER KEY

REVIEW LAB ANSWER KEY REVIEW LAB ANSWER KEY. Witout using SN, find te derivative of eac of te following (you do not need to simplify your answers): a. f x 3x 3 5x x 6 f x 3 3x 5 x 0 b. g x 4 x x x notice te trick ere! x x g

More information

7 Semiparametric Methods and Partially Linear Regression

7 Semiparametric Methods and Partially Linear Regression 7 Semiparametric Metods and Partially Linear Regression 7. Overview A model is called semiparametric if it is described by and were is nite-dimensional (e.g. parametric) and is in nite-dimensional (nonparametric).

More information

The derivative function

The derivative function Roberto s Notes on Differential Calculus Capter : Definition of derivative Section Te derivative function Wat you need to know already: f is at a point on its grap and ow to compute it. Wat te derivative

More information

Preface. Here are a couple of warnings to my students who may be here to get a copy of what happened on a day that you missed.

Preface. Here are a couple of warnings to my students who may be here to get a copy of what happened on a day that you missed. Preface Here are my online notes for my course tat I teac ere at Lamar University. Despite te fact tat tese are my class notes, tey sould be accessible to anyone wanting to learn or needing a refreser

More information

Exam 1 Review Solutions

Exam 1 Review Solutions Exam Review Solutions Please also review te old quizzes, and be sure tat you understand te omework problems. General notes: () Always give an algebraic reason for your answer (graps are not sufficient),

More information

How to Find the Derivative of a Function: Calculus 1

How to Find the Derivative of a Function: Calculus 1 Introduction How to Find te Derivative of a Function: Calculus 1 Calculus is not an easy matematics course Te fact tat you ave enrolled in suc a difficult subject indicates tat you are interested in te

More information

Poisson Equation in Sobolev Spaces

Poisson Equation in Sobolev Spaces Poisson Equation in Sobolev Spaces OcMountain Dayligt Time. 6, 011 Today we discuss te Poisson equation in Sobolev spaces. It s existence, uniqueness, and regularity. Weak Solution. u = f in, u = g on

More information

Consider a function f we ll specify which assumptions we need to make about it in a minute. Let us reformulate the integral. 1 f(x) dx.

Consider a function f we ll specify which assumptions we need to make about it in a minute. Let us reformulate the integral. 1 f(x) dx. Capter 2 Integrals as sums and derivatives as differences We now switc to te simplest metods for integrating or differentiating a function from its function samples. A careful study of Taylor expansions

More information

Polynomial Interpolation

Polynomial Interpolation Capter 4 Polynomial Interpolation In tis capter, we consider te important problem of approximatinga function fx, wose values at a set of distinct points x, x, x,, x n are known, by a polynomial P x suc

More information

Bootstrap confidence intervals in nonparametric regression without an additive model

Bootstrap confidence intervals in nonparametric regression without an additive model Bootstrap confidence intervals in nonparametric regression witout an additive model Dimitris N. Politis Abstract Te problem of confidence interval construction in nonparametric regression via te bootstrap

More information

Polynomials 3: Powers of x 0 + h

Polynomials 3: Powers of x 0 + h near small binomial Capter 17 Polynomials 3: Powers of + Wile it is easy to compute wit powers of a counting-numerator, it is a lot more difficult to compute wit powers of a decimal-numerator. EXAMPLE

More information

2.8 The Derivative as a Function

2.8 The Derivative as a Function .8 Te Derivative as a Function Typically, we can find te derivative of a function f at many points of its domain: Definition. Suppose tat f is a function wic is differentiable at every point of an open

More information

Copyright c 2008 Kevin Long

Copyright c 2008 Kevin Long Lecture 4 Numerical solution of initial value problems Te metods you ve learned so far ave obtained closed-form solutions to initial value problems. A closedform solution is an explicit algebriac formula

More information

Practice Problem Solutions: Exam 1

Practice Problem Solutions: Exam 1 Practice Problem Solutions: Exam 1 1. (a) Algebraic Solution: Te largest term in te numerator is 3x 2, wile te largest term in te denominator is 5x 2 3x 2 + 5. Tus lim x 5x 2 2x 3x 2 x 5x 2 = 3 5 Numerical

More information

HOMEWORK HELP 2 FOR MATH 151

HOMEWORK HELP 2 FOR MATH 151 HOMEWORK HELP 2 FOR MATH 151 Here we go; te second round of omework elp. If tere are oters you would like to see, let me know! 2.4, 43 and 44 At wat points are te functions f(x) and g(x) = xf(x)continuous,

More information

3.1 Extreme Values of a Function

3.1 Extreme Values of a Function .1 Etreme Values of a Function Section.1 Notes Page 1 One application of te derivative is finding minimum and maimum values off a grap. In precalculus we were only able to do tis wit quadratics by find

More information

Function Composition and Chain Rules

Function Composition and Chain Rules Function Composition and s James K. Peterson Department of Biological Sciences and Department of Matematical Sciences Clemson University Marc 8, 2017 Outline 1 Function Composition and Continuity 2 Function

More information

NUMERICAL DIFFERENTIATION. James T. Smith San Francisco State University. In calculus classes, you compute derivatives algebraically: for example,

NUMERICAL DIFFERENTIATION. James T. Smith San Francisco State University. In calculus classes, you compute derivatives algebraically: for example, NUMERICAL DIFFERENTIATION James T Smit San Francisco State University In calculus classes, you compute derivatives algebraically: for example, f( x) = x + x f ( x) = x x Tis tecnique requires your knowing

More information

5.1 We will begin this section with the definition of a rational expression. We

5.1 We will begin this section with the definition of a rational expression. We Basic Properties and Reducing to Lowest Terms 5.1 We will begin tis section wit te definition of a rational epression. We will ten state te two basic properties associated wit rational epressions and go

More information

Cubic Functions: Local Analysis

Cubic Functions: Local Analysis Cubic function cubing coefficient Capter 13 Cubic Functions: Local Analysis Input-Output Pairs, 378 Normalized Input-Output Rule, 380 Local I-O Rule Near, 382 Local Grap Near, 384 Types of Local Graps

More information

Differential Calculus (The basics) Prepared by Mr. C. Hull

Differential Calculus (The basics) Prepared by Mr. C. Hull Differential Calculus Te basics) A : Limits In tis work on limits, we will deal only wit functions i.e. tose relationsips in wic an input variable ) defines a unique output variable y). Wen we work wit

More information

Lecture 15. Interpolation II. 2 Piecewise polynomial interpolation Hermite splines

Lecture 15. Interpolation II. 2 Piecewise polynomial interpolation Hermite splines Lecture 5 Interpolation II Introduction In te previous lecture we focused primarily on polynomial interpolation of a set of n points. A difficulty we observed is tat wen n is large, our polynomial as to

More information

Financial Econometrics Prof. Massimo Guidolin

Financial Econometrics Prof. Massimo Guidolin CLEFIN A.A. 2010/2011 Financial Econometrics Prof. Massimo Guidolin A Quick Review of Basic Estimation Metods 1. Were te OLS World Ends... Consider two time series 1: = { 1 2 } and 1: = { 1 2 }. At tis

More information

Pre-Calculus Review Preemptive Strike

Pre-Calculus Review Preemptive Strike Pre-Calculus Review Preemptive Strike Attaced are some notes and one assignment wit tree parts. Tese are due on te day tat we start te pre-calculus review. I strongly suggest reading troug te notes torougly

More information

2.3 Product and Quotient Rules

2.3 Product and Quotient Rules .3. PRODUCT AND QUOTIENT RULES 75.3 Product and Quotient Rules.3.1 Product rule Suppose tat f and g are two di erentiable functions. Ten ( g (x)) 0 = f 0 (x) g (x) + g 0 (x) See.3.5 on page 77 for a proof.

More information

Polynomial Interpolation

Polynomial Interpolation Capter 4 Polynomial Interpolation In tis capter, we consider te important problem of approximating a function f(x, wose values at a set of distinct points x, x, x 2,,x n are known, by a polynomial P (x

More information

ALGEBRA AND TRIGONOMETRY REVIEW by Dr TEBOU, FIU. A. Fundamental identities Throughout this section, a and b denotes arbitrary real numbers.

ALGEBRA AND TRIGONOMETRY REVIEW by Dr TEBOU, FIU. A. Fundamental identities Throughout this section, a and b denotes arbitrary real numbers. ALGEBRA AND TRIGONOMETRY REVIEW by Dr TEBOU, FIU A. Fundamental identities Trougout tis section, a and b denotes arbitrary real numbers. i) Square of a sum: (a+b) =a +ab+b ii) Square of a difference: (a-b)

More information

IEOR 165 Lecture 10 Distribution Estimation

IEOR 165 Lecture 10 Distribution Estimation IEOR 165 Lecture 10 Distribution Estimation 1 Motivating Problem Consider a situation were we ave iid data x i from some unknown distribution. One problem of interest is estimating te distribution tat

More information

THE IDEA OF DIFFERENTIABILITY FOR FUNCTIONS OF SEVERAL VARIABLES Math 225

THE IDEA OF DIFFERENTIABILITY FOR FUNCTIONS OF SEVERAL VARIABLES Math 225 THE IDEA OF DIFFERENTIABILITY FOR FUNCTIONS OF SEVERAL VARIABLES Mat 225 As we ave seen, te definition of derivative for a Mat 111 function g : R R and for acurveγ : R E n are te same, except for interpretation:

More information

Lecture XVII. Abstract We introduce the concept of directional derivative of a scalar function and discuss its relation with the gradient operator.

Lecture XVII. Abstract We introduce the concept of directional derivative of a scalar function and discuss its relation with the gradient operator. Lecture XVII Abstract We introduce te concept of directional derivative of a scalar function and discuss its relation wit te gradient operator. Directional derivative and gradient Te directional derivative

More information

Time (hours) Morphine sulfate (mg)

Time (hours) Morphine sulfate (mg) Mat Xa Fall 2002 Review Notes Limits and Definition of Derivative Important Information: 1 According to te most recent information from te Registrar, te Xa final exam will be eld from 9:15 am to 12:15

More information

1 The concept of limits (p.217 p.229, p.242 p.249, p.255 p.256) 1.1 Limits Consider the function determined by the formula 3. x since at this point

1 The concept of limits (p.217 p.229, p.242 p.249, p.255 p.256) 1.1 Limits Consider the function determined by the formula 3. x since at this point MA00 Capter 6 Calculus and Basic Linear Algebra I Limits, Continuity and Differentiability Te concept of its (p.7 p.9, p.4 p.49, p.55 p.56). Limits Consider te function determined by te formula f Note

More information

Combining functions: algebraic methods

Combining functions: algebraic methods Combining functions: algebraic metods Functions can be added, subtracted, multiplied, divided, and raised to a power, just like numbers or algebra expressions. If f(x) = x 2 and g(x) = x + 2, clearly f(x)

More information

Order of Accuracy. ũ h u Ch p, (1)

Order of Accuracy. ũ h u Ch p, (1) Order of Accuracy 1 Terminology We consider a numerical approximation of an exact value u. Te approximation depends on a small parameter, wic can be for instance te grid size or time step in a numerical

More information

Chapter 1D - Rational Expressions

Chapter 1D - Rational Expressions - Capter 1D Capter 1D - Rational Expressions Definition of a Rational Expression A rational expression is te quotient of two polynomials. (Recall: A function px is a polynomial in x of degree n, if tere

More information

Local Orthogonal Polynomial Expansion (LOrPE) for Density Estimation

Local Orthogonal Polynomial Expansion (LOrPE) for Density Estimation Local Ortogonal Polynomial Expansion (LOrPE) for Density Estimation Alex Trindade Dept. of Matematics & Statistics, Texas Tec University Igor Volobouev, Texas Tec University (Pysics Dept.) D.P. Amali Dassanayake,

More information

Name: Answer Key No calculators. Show your work! 1. (21 points) All answers should either be,, a (finite) real number, or DNE ( does not exist ).

Name: Answer Key No calculators. Show your work! 1. (21 points) All answers should either be,, a (finite) real number, or DNE ( does not exist ). Mat - Final Exam August 3 rd, Name: Answer Key No calculators. Sow your work!. points) All answers sould eiter be,, a finite) real number, or DNE does not exist ). a) Use te grap of te function to evaluate

More information

MATH1131/1141 Calculus Test S1 v8a

MATH1131/1141 Calculus Test S1 v8a MATH/ Calculus Test 8 S v8a October, 7 Tese solutions were written by Joann Blanco, typed by Brendan Trin and edited by Mattew Yan and Henderson Ko Please be etical wit tis resource It is for te use of

More information

Homework 1 Due: Wednesday, September 28, 2016

Homework 1 Due: Wednesday, September 28, 2016 0-704 Information Processing and Learning Fall 06 Homework Due: Wednesday, September 8, 06 Notes: For positive integers k, [k] := {,..., k} denotes te set of te first k positive integers. Wen p and Y q

More information

LIMITATIONS OF EULER S METHOD FOR NUMERICAL INTEGRATION

LIMITATIONS OF EULER S METHOD FOR NUMERICAL INTEGRATION LIMITATIONS OF EULER S METHOD FOR NUMERICAL INTEGRATION LAURA EVANS.. Introduction Not all differential equations can be explicitly solved for y. Tis can be problematic if we need to know te value of y

More information

Click here to see an animation of the derivative

Click here to see an animation of the derivative Differentiation Massoud Malek Derivative Te concept of derivative is at te core of Calculus; It is a very powerful tool for understanding te beavior of matematical functions. It allows us to optimize functions,

More information

2.3 More Differentiation Patterns

2.3 More Differentiation Patterns 144 te derivative 2.3 More Differentiation Patterns Polynomials are very useful, but tey are not te only functions we need. Tis section uses te ideas of te two previous sections to develop tecniques for

More information

NADARAYA WATSON ESTIMATE JAN 10, 2006: version 2. Y ik ( x i

NADARAYA WATSON ESTIMATE JAN 10, 2006: version 2. Y ik ( x i NADARAYA WATSON ESTIMATE JAN 0, 2006: version 2 DATA: (x i, Y i, i =,..., n. ESTIMATE E(Y x = m(x by n i= ˆm (x = Y ik ( x i x n i= K ( x i x EXAMPLES OF K: K(u = I{ u c} (uniform or box kernel K(u = u

More information

Sin, Cos and All That

Sin, Cos and All That Sin, Cos and All Tat James K. Peterson Department of Biological Sciences and Department of Matematical Sciences Clemson University Marc 9, 2017 Outline Sin, Cos and all tat! A New Power Rule Derivatives

More information

Differentiation in higher dimensions

Differentiation in higher dimensions Capter 2 Differentiation in iger dimensions 2.1 Te Total Derivative Recall tat if f : R R is a 1-variable function, and a R, we say tat f is differentiable at x = a if and only if te ratio f(a+) f(a) tends

More information

Derivatives of Exponentials

Derivatives of Exponentials mat 0 more on derivatives: day 0 Derivatives of Eponentials Recall tat DEFINITION... An eponential function as te form f () =a, were te base is a real number a > 0. Te domain of an eponential function

More information

Exercises for numerical differentiation. Øyvind Ryan

Exercises for numerical differentiation. Øyvind Ryan Exercises for numerical differentiation Øyvind Ryan February 25, 2013 1. Mark eac of te following statements as true or false. a. Wen we use te approximation f (a) (f (a +) f (a))/ on a computer, we can

More information

Continuity. Example 1

Continuity. Example 1 Continuity MATH 1003 Calculus and Linear Algebra (Lecture 13.5) Maoseng Xiong Department of Matematics, HKUST A function f : (a, b) R is continuous at a point c (a, b) if 1. x c f (x) exists, 2. f (c)

More information

LIMITS AND DERIVATIVES CONDITIONS FOR THE EXISTENCE OF A LIMIT

LIMITS AND DERIVATIVES CONDITIONS FOR THE EXISTENCE OF A LIMIT LIMITS AND DERIVATIVES Te limit of a function is defined as te value of y tat te curve approaces, as x approaces a particular value. Te limit of f (x) as x approaces a is written as f (x) approaces, as

More information

A = h w (1) Error Analysis Physics 141

A = h w (1) Error Analysis Physics 141 Introduction In all brances of pysical science and engineering one deals constantly wit numbers wic results more or less directly from experimental observations. Experimental observations always ave inaccuracies.

More information

2.1 THE DEFINITION OF DERIVATIVE

2.1 THE DEFINITION OF DERIVATIVE 2.1 Te Derivative Contemporary Calculus 2.1 THE DEFINITION OF DERIVATIVE 1 Te grapical idea of a slope of a tangent line is very useful, but for some uses we need a more algebraic definition of te derivative

More information

Lab 6 Derivatives and Mutant Bacteria

Lab 6 Derivatives and Mutant Bacteria Lab 6 Derivatives and Mutant Bacteria Date: September 27, 20 Assignment Due Date: October 4, 20 Goal: In tis lab you will furter explore te concept of a derivative using R. You will use your knowledge

More information

Notes on wavefunctions II: momentum wavefunctions

Notes on wavefunctions II: momentum wavefunctions Notes on wavefunctions II: momentum wavefunctions and uncertainty Te state of a particle at any time is described by a wavefunction ψ(x). Tese wavefunction must cange wit time, since we know tat particles

More information

Lesson 6: The Derivative

Lesson 6: The Derivative Lesson 6: Te Derivative Def. A difference quotient for a function as te form f(x + ) f(x) (x + ) x f(x + x) f(x) (x + x) x f(a + ) f(a) (a + ) a Notice tat a difference quotient always as te form of cange

More information

Kernel Density Estimation

Kernel Density Estimation Kernel Density Estimation Univariate Density Estimation Suppose tat we ave a random sample of data X 1,..., X n from an unknown continuous distribution wit probability density function (pdf) f(x) and cumulative

More information

Precalculus Test 2 Practice Questions Page 1. Note: You can expect other types of questions on the test than the ones presented here!

Precalculus Test 2 Practice Questions Page 1. Note: You can expect other types of questions on the test than the ones presented here! Precalculus Test 2 Practice Questions Page Note: You can expect oter types of questions on te test tan te ones presented ere! Questions Example. Find te vertex of te quadratic f(x) = 4x 2 x. Example 2.

More information

1. Consider the trigonometric function f(t) whose graph is shown below. Write down a possible formula for f(t).

1. Consider the trigonometric function f(t) whose graph is shown below. Write down a possible formula for f(t). . Consider te trigonometric function f(t) wose grap is sown below. Write down a possible formula for f(t). Tis function appears to be an odd, periodic function tat as been sifted upwards, so we will use

More information

3.4 Worksheet: Proof of the Chain Rule NAME

3.4 Worksheet: Proof of the Chain Rule NAME Mat 1170 3.4 Workseet: Proof of te Cain Rule NAME Te Cain Rule So far we are able to differentiate all types of functions. For example: polynomials, rational, root, and trigonometric functions. We are

More information

3.4 Algebraic Limits. Ex 1) lim. Ex 2)

3.4 Algebraic Limits. Ex 1) lim. Ex 2) Calculus Maimus.4 Algebraic Limits At tis point, you sould be very comfortable finding its bot grapically and numerically wit te elp of your graping calculator. Now it s time to practice finding its witout

More information

Solution. Solution. f (x) = (cos x)2 cos(2x) 2 sin(2x) 2 cos x ( sin x) (cos x) 4. f (π/4) = ( 2/2) ( 2/2) ( 2/2) ( 2/2) 4.

Solution. Solution. f (x) = (cos x)2 cos(2x) 2 sin(2x) 2 cos x ( sin x) (cos x) 4. f (π/4) = ( 2/2) ( 2/2) ( 2/2) ( 2/2) 4. December 09, 20 Calculus PracticeTest s Name: (4 points) Find te absolute extrema of f(x) = x 3 0 on te interval [0, 4] Te derivative of f(x) is f (x) = 3x 2, wic is zero only at x = 0 Tus we only need

More information

Recall from our discussion of continuity in lecture a function is continuous at a point x = a if and only if

Recall from our discussion of continuity in lecture a function is continuous at a point x = a if and only if Computational Aspects of its. Keeping te simple simple. Recall by elementary functions we mean :Polynomials (including linear and quadratic equations) Eponentials Logaritms Trig Functions Rational Functions

More information

. If lim. x 2 x 1. f(x+h) f(x)

. If lim. x 2 x 1. f(x+h) f(x) Review of Differential Calculus Wen te value of one variable y is uniquely determined by te value of anoter variable x, ten te relationsip between x and y is described by a function f tat assigns a value

More information

SECTION 3.2: DERIVATIVE FUNCTIONS and DIFFERENTIABILITY

SECTION 3.2: DERIVATIVE FUNCTIONS and DIFFERENTIABILITY (Section 3.2: Derivative Functions and Differentiability) 3.2.1 SECTION 3.2: DERIVATIVE FUNCTIONS and DIFFERENTIABILITY LEARNING OBJECTIVES Know, understand, and apply te Limit Definition of te Derivative

More information

SECTION 1.10: DIFFERENCE QUOTIENTS LEARNING OBJECTIVES

SECTION 1.10: DIFFERENCE QUOTIENTS LEARNING OBJECTIVES (Section.0: Difference Quotients).0. SECTION.0: DIFFERENCE QUOTIENTS LEARNING OBJECTIVES Define average rate of cange (and average velocity) algebraically and grapically. Be able to identify, construct,

More information

Excursions in Computing Science: Week v Milli-micro-nano-..math Part II

Excursions in Computing Science: Week v Milli-micro-nano-..math Part II Excursions in Computing Science: Week v Milli-micro-nano-..mat Part II T. H. Merrett McGill University, Montreal, Canada June, 5 I. Prefatory Notes. Cube root of 8. Almost every calculator as a square-root

More information

Continuity and Differentiability Worksheet

Continuity and Differentiability Worksheet Continuity and Differentiability Workseet (Be sure tat you can also do te grapical eercises from te tet- Tese were not included below! Typical problems are like problems -3, p. 6; -3, p. 7; 33-34, p. 7;

More information

Math 212-Lecture 9. For a single-variable function z = f(x), the derivative is f (x) = lim h 0

Math 212-Lecture 9. For a single-variable function z = f(x), the derivative is f (x) = lim h 0 3.4: Partial Derivatives Definition Mat 22-Lecture 9 For a single-variable function z = f(x), te derivative is f (x) = lim 0 f(x+) f(x). For a function z = f(x, y) of two variables, to define te derivatives,

More information

Solution for the Homework 4

Solution for the Homework 4 Solution for te Homework 4 Problem 6.5: In tis section we computed te single-particle translational partition function, tr, by summing over all definite-energy wavefunctions. An alternative approac, owever,

More information

2.11 That s So Derivative

2.11 That s So Derivative 2.11 Tat s So Derivative Introduction to Differential Calculus Just as one defines instantaneous velocity in terms of average velocity, we now define te instantaneous rate of cange of a function at a point

More information

Chapter 2 Limits and Continuity

Chapter 2 Limits and Continuity 4 Section. Capter Limits and Continuity Section. Rates of Cange and Limits (pp. 6) Quick Review.. f () ( ) () 4 0. f () 4( ) 4. f () sin sin 0 4. f (). 4 4 4 6. c c c 7. 8. c d d c d d c d c 9. 8 ( )(

More information

MAT244 - Ordinary Di erential Equations - Summer 2016 Assignment 2 Due: July 20, 2016

MAT244 - Ordinary Di erential Equations - Summer 2016 Assignment 2 Due: July 20, 2016 MAT244 - Ordinary Di erential Equations - Summer 206 Assignment 2 Due: July 20, 206 Full Name: Student #: Last First Indicate wic Tutorial Section you attend by filling in te appropriate circle: Tut 0

More information

LECTURE 14 NUMERICAL INTEGRATION. Find

LECTURE 14 NUMERICAL INTEGRATION. Find LECTURE 14 NUMERCAL NTEGRATON Find b a fxdx or b a vx ux fx ydy dx Often integration is required. However te form of fx may be suc tat analytical integration would be very difficult or impossible. Use

More information

Kernel Density Based Linear Regression Estimate

Kernel Density Based Linear Regression Estimate Kernel Density Based Linear Regression Estimate Weixin Yao and Zibiao Zao Abstract For linear regression models wit non-normally distributed errors, te least squares estimate (LSE will lose some efficiency

More information

POLYNOMIAL AND SPLINE ESTIMATORS OF THE DISTRIBUTION FUNCTION WITH PRESCRIBED ACCURACY

POLYNOMIAL AND SPLINE ESTIMATORS OF THE DISTRIBUTION FUNCTION WITH PRESCRIBED ACCURACY APPLICATIONES MATHEMATICAE 36, (29), pp. 2 Zbigniew Ciesielski (Sopot) Ryszard Zieliński (Warszawa) POLYNOMIAL AND SPLINE ESTIMATORS OF THE DISTRIBUTION FUNCTION WITH PRESCRIBED ACCURACY Abstract. Dvoretzky

More information

Section 2.7 Derivatives and Rates of Change Part II Section 2.8 The Derivative as a Function. at the point a, to be. = at time t = a is

Section 2.7 Derivatives and Rates of Change Part II Section 2.8 The Derivative as a Function. at the point a, to be. = at time t = a is Mat 180 www.timetodare.com Section.7 Derivatives and Rates of Cange Part II Section.8 Te Derivative as a Function Derivatives ( ) In te previous section we defined te slope of te tangent to a curve wit

More information

5 Ordinary Differential Equations: Finite Difference Methods for Boundary Problems

5 Ordinary Differential Equations: Finite Difference Methods for Boundary Problems 5 Ordinary Differential Equations: Finite Difference Metods for Boundary Problems Read sections 10.1, 10.2, 10.4 Review questions 10.1 10.4, 10.8 10.9, 10.13 5.1 Introduction In te previous capters we

More information

Section 2.1 The Definition of the Derivative. We are interested in finding the slope of the tangent line at a specific point.

Section 2.1 The Definition of the Derivative. We are interested in finding the slope of the tangent line at a specific point. Popper 6: Review of skills: Find tis difference quotient. f ( x ) f ( x) if f ( x) x Answer coices given in audio on te video. Section.1 Te Definition of te Derivative We are interested in finding te slope

More information

Physically Based Modeling: Principles and Practice Implicit Methods for Differential Equations

Physically Based Modeling: Principles and Practice Implicit Methods for Differential Equations Pysically Based Modeling: Principles and Practice Implicit Metods for Differential Equations David Baraff Robotics Institute Carnegie Mellon University Please note: Tis document is 997 by David Baraff

More information

, meant to remind us of the definition of f (x) as the limit of difference quotients: = lim

, meant to remind us of the definition of f (x) as the limit of difference quotients: = lim Mat 132 Differentiation Formulas Stewart 2.3 So far, we ave seen ow various real-world problems rate of cange and geometric problems tangent lines lead to derivatives. In tis section, we will see ow to

More information

Section 3: The Derivative Definition of the Derivative

Section 3: The Derivative Definition of the Derivative Capter 2 Te Derivative Business Calculus 85 Section 3: Te Derivative Definition of te Derivative Returning to te tangent slope problem from te first section, let's look at te problem of finding te slope

More information

Taylor Series and the Mean Value Theorem of Derivatives

Taylor Series and the Mean Value Theorem of Derivatives 1 - Taylor Series and te Mean Value Teorem o Derivatives Te numerical solution o engineering and scientiic problems described by matematical models oten requires solving dierential equations. Dierential

More information

MVT and Rolle s Theorem

MVT and Rolle s Theorem AP Calculus CHAPTER 4 WORKSHEET APPLICATIONS OF DIFFERENTIATION MVT and Rolle s Teorem Name Seat # Date UNLESS INDICATED, DO NOT USE YOUR CALCULATOR FOR ANY OF THESE QUESTIONS In problems 1 and, state

More information

2.3 Algebraic approach to limits

2.3 Algebraic approach to limits CHAPTER 2. LIMITS 32 2.3 Algebraic approac to its Now we start to learn ow to find its algebraically. Tis starts wit te simplest possible its, and ten builds tese up to more complicated examples. Fact.

More information

Applications of the van Trees inequality to non-parametric estimation.

Applications of the van Trees inequality to non-parametric estimation. Brno-06, Lecture 2, 16.05.06 D/Stat/Brno-06/2.tex www.mast.queensu.ca/ blevit/ Applications of te van Trees inequality to non-parametric estimation. Regular non-parametric problems. As an example of suc

More information

0.1 Differentiation Rules

0.1 Differentiation Rules 0.1 Differentiation Rules From our previous work we ve seen tat it can be quite a task to calculate te erivative of an arbitrary function. Just working wit a secon-orer polynomial tings get pretty complicate

More information

Phase space in classical physics

Phase space in classical physics Pase space in classical pysics Quantum mecanically, we can actually COU te number of microstates consistent wit a given macrostate, specified (for example) by te total energy. In general, eac microstate

More information

1 1. Rationalize the denominator and fully simplify the radical expression 3 3. Solution: = 1 = 3 3 = 2

1 1. Rationalize the denominator and fully simplify the radical expression 3 3. Solution: = 1 = 3 3 = 2 MTH - Spring 04 Exam Review (Solutions) Exam : February 5t 6:00-7:0 Tis exam review contains questions similar to tose you sould expect to see on Exam. Te questions included in tis review, owever, are

More information

Chapter 2 Ising Model for Ferromagnetism

Chapter 2 Ising Model for Ferromagnetism Capter Ising Model for Ferromagnetism Abstract Tis capter presents te Ising model for ferromagnetism, wic is a standard simple model of a pase transition. Using te approximation of mean-field teory, te

More information

Math 102 TEST CHAPTERS 3 & 4 Solutions & Comments Fall 2006

Math 102 TEST CHAPTERS 3 & 4 Solutions & Comments Fall 2006 Mat 102 TEST CHAPTERS 3 & 4 Solutions & Comments Fall 2006 f(x+) f(x) 10 1. For f(x) = x 2 + 2x 5, find ))))))))) and simplify completely. NOTE: **f(x+) is NOT f(x)+! f(x+) f(x) (x+) 2 + 2(x+) 5 ( x 2

More information

THE ROYAL STATISTICAL SOCIETY GRADUATE DIPLOMA EXAMINATION MODULE 5

THE ROYAL STATISTICAL SOCIETY GRADUATE DIPLOMA EXAMINATION MODULE 5 THE ROYAL STATISTICAL SOCIETY GRADUATE DIPLOMA EXAMINATION NEW MODULAR SCHEME introduced from te examinations in 009 MODULE 5 SOLUTIONS FOR SPECIMEN PAPER B THE QUESTIONS ARE CONTAINED IN A SEPARATE FILE

More information

Bootstrap prediction intervals for Markov processes

Bootstrap prediction intervals for Markov processes arxiv: arxiv:0000.0000 Bootstrap prediction intervals for Markov processes Li Pan and Dimitris N. Politis Li Pan Department of Matematics University of California San Diego La Jolla, CA 92093-0112, USA

More information

4. The slope of the line 2x 7y = 8 is (a) 2/7 (b) 7/2 (c) 2 (d) 2/7 (e) None of these.

4. The slope of the line 2x 7y = 8 is (a) 2/7 (b) 7/2 (c) 2 (d) 2/7 (e) None of these. Mat 11. Test Form N Fall 016 Name. Instructions. Te first eleven problems are wort points eac. Te last six problems are wort 5 points eac. For te last six problems, you must use relevant metods of algebra

More information

Section 2: The Derivative Definition of the Derivative

Section 2: The Derivative Definition of the Derivative Capter 2 Te Derivative Applied Calculus 80 Section 2: Te Derivative Definition of te Derivative Suppose we drop a tomato from te top of a 00 foot building and time its fall. Time (sec) Heigt (ft) 0.0 00

More information

Some Review Problems for First Midterm Mathematics 1300, Calculus 1

Some Review Problems for First Midterm Mathematics 1300, Calculus 1 Some Review Problems for First Midterm Matematics 00, Calculus. Consider te trigonometric function f(t) wose grap is sown below. Write down a possible formula for f(t). Tis function appears to be an odd,

More information

Math 312 Lecture Notes Modeling

Math 312 Lecture Notes Modeling Mat 3 Lecture Notes Modeling Warren Weckesser Department of Matematics Colgate University 5 7 January 006 Classifying Matematical Models An Example We consider te following scenario. During a storm, a

More information