OPTIMAL PREDICTION UNDER ASYMMETRIC LOSS 1. By Peter F. Christoffersen and Francis X. Diebold 2 1. INTRODUCTION
|
|
- Melvyn Craig
- 6 years ago
- Views:
Transcription
1 OPTIMAL PREDICTION UNDER ASYMMETRIC LOSS 1 By Peter F. Cristoffersen and Francis X. Diebold Cristoffersen P. and Diebold F.X. (1997) "Optimal Prediction Under Asymmetric Loss" Econometric Teory Keywords: Forecasting loss function asymmetric loss nonlinear eteroskedasticity 1. INTRODUCTION A MOMENT'S REFLECTION yields te insigt tat prediction problems involving asymmetric loss structures arise routinely as a myriad of situation-specific factors may render positive errors more (or less) costly tan negative errors. Te potential necessity of allowing for asymmetric loss as long been acknowledged. Granger and Newbold (1986) for example note tat altoug "an assumption of symmetry about te conditional mean... is likely to be an easy one to accept... an assumption of symmetry for te cost function is muc less acceptable" (p. 15). Practitioners routinely eco tis sentiment (e.g. Stockman 1987). In tis paper we treat te prediction problem under general loss structures building on te classic work of Granger (1969). In Section we caracterize te optimal predictor for non- Gaussian processes under asymmetric loss. Te results apply for example to important classes of conditionally eteroskedastic processes. In Section 3 we provide analytic solutions for te optimal predictor under two popular analytically-tractable asymmetric loss functions. In Section 4 we provide metods for approximating te optimal predictor under more general loss functions. We conclude in Section 5.. OPTIMAL PREDICTION FOR NON-GAUSSIAN PROCESSES Granger (1969) studies Gaussian processes and sows tat under asymmetric loss te optimal predictor is te conditional mean plus a constant bias term. Granger's fundamental result owever as two key limitations. First te Gaussian assumption implies a constant
2 -- conditional prediction-error variance. Tis is unfortunate because conditional eteroskedasticity is widespread in economic and financial data. Second te loss function must be of predictionerror form; tat is L(y t ) L(y t ) L(e t ) were y t is te -step-aead realization is te -step-aead forecast (made at time t) and is te corresponding forecast error. More general functions of realizations and predictions are excluded. Let us begin ten by generalizing Granger's result to allow for conditional variance dynamics. We acieve tis most simply by working in a conditionally-gaussian but not necessarily unconditionally-gaussian environment wit prediction-error loss. Subsequently we sall allow for bot conditional non-normality and more general loss functions. e t PROPOSITION 1: If y N(µ ) is a conditionally Gaussian process and L(e t ) is any loss function defined on te -step-aead prediction error e t ten te optimal predictor is of te form µ were depends only on te loss function and te conditional prediction-error variance var(y ) var(e ). PROOF: See Appendix. Te optimal predictor under conditional normality is not necessarily just a constant added to te conditional mean because te conditional prediction-error variance may be time-varying. Conditionally Gaussian GARCH processes for example fall under te jurisdiction of Proposition 1. Tus under asymmetric loss conditional variance dynamics are important not only for interval prediction but also for point prediction. If loss is asymmetric but conditional eteroskedasticity is ignored te resulting point predictions will be suboptimal and may ave dramatically greater conditionally expected loss in consequence. Te result of Proposition 1 tat te "adjustment factor" depends only on te conditional variance depends crucially on conditional normality. We can dispense wit conditional
3 -3- normality and still obtain a sarp result owever wic is a straigtforward extension of Proposition 1. PROPOSITION : If y as conditional mean µ and a vector of (possibly time varying) conditional moments of order two and iger and L(e t ) is any loss function t+ t defined on te -step-aead prediction error e t ten te optimal predictor is of te form µ were depends only on te loss function and t+ t. PROOF: See Appendix. Note owever tat altoug Propostion does not require a Gaussian process it does require prediction-error loss. In Section 4 we will relax tat assumption as well. 3. ANALYTIC SOLUTIONS UNDER LINEX AND LINLIN LOSS Here we examine two asymmetric loss functions ("linex" and "linlin") for wic it is possible to solve analytically for te optimal predictor. To maintain continuity of exposition we work trougout tis section wit te conditionally Gaussian process y N(µ ). 3 For eac loss function we caracterize te optimal predictor µ and we compare its conditionally expected loss to tat of two competitors te conditional mean µ and te pseudo-optimal predictor µ were depends only on te loss function and te unconditional prediction-error variance var(e t ). Te optimal predictor acknowledges loss asymmetry and te possibility of conditional eteroskedasticity troug a possibly time-varying adjustment to te conditional mean. Te conditional mean in contrast is always suboptimal as it incorporates no adjustment. Te pseudo-optimal predictor is intermediate in tat it incorporates only a constant adjustment for asymmetry; tus it is fully optimal only in te conditionally omoskedastic case t Linex Loss
4 -4- Te "linex" loss function introduced by Varian (1974) and used by Zellner (1986) is L(x) b exp(ax) ax 1 a {0} b. It is so-named because wen a>0 loss is approximately linear to te left of te origin and approximately exponential to te rigt and conversely wen a<0. Te optimal -step-aead predictor under linex loss solves min E t b exp(a(y t )) a(y t ) 1. Differentiating and using te conditional moment-generating function for a conditionally a Gaussian variate we obtain µ. Similar calculations reveal tat te pseudooptimal predictor is were a µ var(e t ) is te unconditional -step-aead prediction-error variance. Proposition 1 sows tat te optimal predictor under conditional normality is te conditional mean plus a function of te conditional prediction-error variance. Under linex loss te function is a simple linear one depending on te degree of asymmetry of te loss function as 4 captured in te parameter a. Te reason is simple--wen a is positive for example positive prediction errors are more devastating tan negative errors so a negative conditionally expected error is desirable. Te optimal amount of bias depends on te conditional prediction-error variance of te process; as it grows so too does te optimal amount of bias in order to avoid large positive prediction errors. Effectively optimal prediction under asymmetric loss corresponds to conditional-mean prediction of a transformed series were te transformation reflects bot te loss function and te iger-order conditional moments of te original series.
5 For example te optimal predictor of y under conditional normality and linex loss t+ µ a -5- x a is te conditional mean of x t+ were t y t. 5 Inserting te optimal pseudo-optimal and conditional mean predictors into te conditionally expected loss expression we see tat te conditionally-expected linex losses are ba / b[exp(a ( )/) a / 1] b[exp(a and /) 1] respectively. By construction te conditionally expected loss of te optimal predictor is less tan or equal to tat of any oter predictor. Interestingly owever it is not possible to rank te pseudo-optimal as superior to te conditional mean predictor. Tedious but straigtforward algebra reveals tat for sufficiently small values of (depending non-linearly on te values of a and ) te conditionally expected loss of te conditional mean will be smaller tan tat of te pseudooptimal predictor. In very low volatility times te conditionally optimal amount of bias is very small resulting in a lower conditionally expected loss for te conditional mean tan for te pseudo-optimal predictor te bias of wic is optimal in "average" times but too low in lowvolatility times. Te situation is illustrated in Figure 1 in wic we plot conditionally expected linex loss as a function of for eac of te tree predictors. Te conditionally expected loss of te optimal predictor is linear in and is of course always lowest. Te losses of te pseudooptimal and te optimal predictors coincide wen 1. As falls below te loss of te conditional mean intersects te loss of te pseudo-optimal predictor from above. As gets close to zero te optimal predictor incorporates progressively smaller corrections to te conditional mean so te conditionally expected losses of te optimal and conditional mean predictors coincide. 3.. Linlin Loss
6 -6- Te "linlin" loss function L(y t ) a y t if (y t ) > 0 b y t if (y t ) 0 so-called because of its linearity on eac side of te origin was used by Granger (1969) and is te loss function underlying quantile regression. Te optimal predictor solves min a (y t )f(y )dy t b (y t )f(y )dy t. Te first-order condition is F( t ) conditional density of y. t+ a(1 F( t )) b F( t ) 0 a a b were F(y ) is te conditional c.d.f. of y and f(y ) t+ is te In te conditionally Gaussian case we ave from Proposition 1 tat wic is equivalent to F( t ) Pr (y t (µ t )) t Pr y t µ t t t a a b were (z) is te N(01) c.d.f. It follows tat te conditionally optimal amount of bias is a b ŷ so tat t µ a b. tat te pseudo-optimal predictor is µ a b. Similar calculations reveal Now let us compute conditionally expected linlin loss for te optimal pseudo-optimal and conditional mean predictors. Recall te formulae for te truncated expectation 6 y t f(y t E t {y t (y t > )} y t f(y )dy t 1 F( t ) E t {y t (y t < )} F( t ) t )dy t
7 and substitute tem into te expected loss expression to obtain -7- {L(y t )} a(1 F( t ))[E t y t (y t > ) ] bf( t )[E t y t (y t < ) But under conditional normality ( E t (y t (y t > )) µ ) 1 ( ) E (y t t (y <ŷ )) µ µ were and ( ) is te N(01) p.d.f. Substituting into te conditionally expected loss expression we obtain (after some algebraic manipulation) ( ) ( ) E t (L(y t )) (a b) ( ) a( t µ ) (a b) ( )( t µ ). For te optimal predictor (a b) a b yielding an expected loss of (a b) a b a b. For te pseudo-optimal predictor For te conditional mean predictor a a b 0 yielding an expected loss of (a b) a b yielding an expected loss of a b a b. E t (L(y t )) (a b) /. Qualitatively te situation is identical to tat sown in Figure 1 for te linex case. 4. APPROXIMATING THE OPTIMAL PREDICTOR Te analytic results above rely on simple loss functions. In general owever it is not possible to solve analytically for te optimal predictor. Here we develop an approximately optimal predictor via series expansions. Te approac is of interest because it frees us from two potentially restrictive assumptions -- conditional normality and prediction-error loss. For te moment maintain te conditional normality assumption and assume tat te optimal predictor exists and is unique G(µ ) were G( ) is at least twice
8 -8- continuously differentiable. Ten we can take a second order Taylor series expansion around te unconditional (and time invariant) moments µ and G(µ µ ) G (µ ) µ 1 (µ µ ) G (µ ) µ µ. Rewrite tis as 0 1 µ 3 (µ ) 4 ( ) 5 (µ ) y t ( ) were ( ) and i H i (µ ) i Because te function G( ) is generally unknown so too are te H( ) functions. But µ and are known and te ˆ N ˆ N minimization tat defines can be done over a very long simulated realization of lengt N argmin N L(y t y t ( )). Under regularity conditions given in te Appendix te t 1 following proposition is immediate. PROPOSITION 3: As N y t (ˆ N) y t ( 0 ) were y t ( 0 ) is te best predictor witin te y t ( ) family wit respect to te metric L( ). PROOF: See Appendix. A number of remarks are in order. First te -step-aead conditional expectation and te corresponding conditional variance may be computed conveniently using te Kalman filter recursions. Second if loss is in fact of prediction-error form L(e ) one may set = nd = t = 0 a priori due to Proposition 1. Tird it is clear tat iger-order expansions in and µ may be entertained and may lead to improvements. Fourt conditional non-normality may be andled wit expansions involving more tan te first two conditional moments (e.g. involving conditional skewness and kurtosis). Fift and related parametric economy can be
9 -9- acieved in conditionally non-gaussian cases using te autoregressive conditional density framework of Hansen (1994). Hansen's framework exploits parametric conditional mean and variance functions but allows for iger-order conditional dynamics by letting te normalized variable z t ( ) y t µ ( ) / ( ) follow a distribution wit possibly time varying "sape" parameters suc as a t-distribution wit time-varying degrees of freedom (and variance standardized to 1). Sixt in bot te conditionally Gaussian and conditionally non-gaussian cases one is of course not limited to series expansions; oter nonparametric functional estimators may be used. 5. SUMMARY AND CONCLUDING REMARKS Tis paper is part of a researc program aimed at allowing for general loss structures in estimation model selection prediction and forecast evaluation. Recently a number of autors ave made progress toward tat goal including Weiss (1994) on estimation Pillips (1994) on model selection and Diebold and Mariano (1995) on forecast evaluation. Here we focused on prediction and analyzed te optimal prediction problem under asymmetric loss. We computed te optimal predictor analytically in two leading tractable cases and sowed ow to compute it numerically in less tractable cases. A key teme is tat te conditionally optimal forecast is biased and tat te conditionally optimal amount of bias is time-varying in general and depends on iger-order conditional moments. Tus even for models wit linear conditional-mean structure te optimal predictor is in general nonlinear tereby providing a link wit te broader nonlinear time series literature. Interestingly some important recent work in dynamic economic teory is very muc linked to te idea of prediction under asymmetric loss discussed ere. Building on Wittle (1990) Hansen Sargent and Tallarini (1993) set up and motivate a general-equilibrium economy
10 -10- wit "risk sensitive" preferences resulting in equilibria wit certainty-equivalence properties. Tus te prediction and decision problems may be done sequentially--but prediction is done wit respect to a distorted probability measure tat yields predictions different from te conditional mean. University of Pennsylvania APPENDIX PROOF OF PROPOSITION 1: We seek te predictor tat solves min E t L(y t ) min L(y t ) f(y ) dy t. (Here and trougout E t (x) denotes E(x t ). ) Witout loss of generality we can write µ and y t µ x t so tat argmin E t L(y t ) µ argmin L(x t ) f(x ) dx t. Because f(x ) depends on but not µ so too does te tat solves te minimization problem depend on but not µ. Q.E.D. PROOF OF PROPOSITION : Precisely parallels tat of Proposition 1. Q.E.D. PROOF OF PROPOSITION 3: Following Amemiya (1985) we require tree conditions: k (1) 0 a compact subset of. N () L N ( ) L(y t y t ( )) is continuous in for all y=(y...y ) and is a t 1 1+ N+
11 uniformly in as N and L( ) attains a unique global minimum at. argmin Under te conditions LN ( ) converges in probability to by te argument of ˆ N -11- measurable function of y for all. -1 (3) N L N( ) converges to a nonstocastic continuous function L( ) in probability Amemiya (1985 p. 107). Tus y t (ˆN) converges in probability to y t ( 0 ) by continuity of 0 0 y t (ˆN). Q.E.D. REFERENCES Amemiya T.: Advanced Econometrics. Cambridge Mass.: Harvard University Press Cristoffersen P.F. and Diebold F.X. (1994): "Optimal Prediction Under Asymmetric Loss" National Bureau of Economic Researc Tecnical Working Paper No. 167 Cambridge Mass. Diebold F.X. and R.S. Mariano (1995): "Comparing Predictive Accuracy" Journal of Business and Economic Statistics Granger C.W.J. (1969): "Prediction wit a Generalized Cost of Error Function" Operational Researc Quarterly Granger C.W.J. and P. Newbold: Forecasting Economic Time Series (Second edition). Orlando: Academic Press Hansen B.E. (1994): "Autoregressive Conditional Density Estimation" International Economic Review Hansen L.P. T.J. Sargent and T.D. Tallarini (1993): "Pessimism Neurosis and Feelings About Risk in General Equilibrium" Manuscript University of Cicago. Pillips P.C.B. (1994): "Bayes Models and Macroeconomic Activity" Manuscript Yale
12 -1- University. Stockman A.C. (1987): "Economic Teory and Excange Rate Forecasts" International Journal of Forecasting Varian H.: "A Bayesian Approac to Real Estate Assessment" in Studies in Bayesian Econometrics and Statistics in Honor of L.J. Savage ed. by S.E. Feinberg and A. Zellner. Amsterdam: Nort-Holland Weiss A.A. (1994): "Estimating Time Series Models Using te Relevant Cost Function" Manuscript Department of Economics University of Soutern California. Wittle P.: Risk-Sensitive Optimal Control. New York: Jon Wiley Zellner A. (1986): "Bayesian Estimation and Prediction Using Asymmetric Loss Functions" Journal of te American Statistical Association
13 -13- FOOTNOTES 1. Tis paper is a eavily-revised and sortened version of parts of Cristoffersen and Diebold (1994) wic may be consulted for additional results discussion and examples.. We benefitted from constructive comments from te Co-Editor and two referees as well as from Clive Granger Hasem Pesaran Enrique Sentana Bob Stine Jim Stock Ken Wallis and numerous conference and seminar participants. Remaining inadequacies are ours alone. We tank te National Science Foundation te Sloan Foundation te University of Pennsylvania Researc Foundation and te Barbara and Edward Netter Fellowsip for support. 3. As will be made clear owever altoug conditional normality is crucial to our derivation of te optimal predictor under linex loss it may readily be discarded under linlin loss. 4. Note tat as a 0 te conditionally optimal amount of bias approaces zero. Quadratic loss obtains as a 0 because if a is small one can replace te exponential part of te loss function by te first two terms of its Taylor series expansion yielding te approximation L(x) x. 5. Because y is conditionally normal wit E[y ] µ x is conditionally normal wit t+ a E[x ] µ ŷ t. 6. Note tat wit linlin loss (in contrast to linex loss) it is very easy even for non-gaussian conditional distributions to find te optimal predictor -- just draw te conditional c.d.f. and read te value on te x-axis corresponding to a/(a+b). More formally is simply te (a/(a+b))t conditional quantile. Wen a=b of course median. t+ F a b t so is te conditional
14 Figure 1 Conditionally Expected Linex Loss of Conditional Mean Pseudo-Optimal and Optimal Predictors Notes to Figure: Te Linex loss parameters are set to a=nd b=. Te unconditional variance is fixed at 1.
The total error in numerical differentiation
AMS 147 Computational Metods and Applications Lecture 08 Copyrigt by Hongyun Wang, UCSC Recap: Loss of accuracy due to numerical cancellation A B 3, 3 ~10 16 In calculating te difference between A and
More informationNumerical Differentiation
Numerical Differentiation Finite Difference Formulas for te first derivative (Using Taylor Expansion tecnique) (section 8.3.) Suppose tat f() = g() is a function of te variable, and tat as 0 te function
More informationFinancial Econometrics Prof. Massimo Guidolin
CLEFIN A.A. 2010/2011 Financial Econometrics Prof. Massimo Guidolin A Quick Review of Basic Estimation Metods 1. Were te OLS World Ends... Consider two time series 1: = { 1 2 } and 1: = { 1 2 }. At tis
More informationA MONTE CARLO ANALYSIS OF THE EFFECTS OF COVARIANCE ON PROPAGATED UNCERTAINTIES
A MONTE CARLO ANALYSIS OF THE EFFECTS OF COVARIANCE ON PROPAGATED UNCERTAINTIES Ronald Ainswort Hart Scientific, American Fork UT, USA ABSTRACT Reports of calibration typically provide total combined uncertainties
More informationLECTURE 14 NUMERICAL INTEGRATION. Find
LECTURE 14 NUMERCAL NTEGRATON Find b a fxdx or b a vx ux fx ydy dx Often integration is required. However te form of fx may be suc tat analytical integration would be very difficult or impossible. Use
More informationlecture 26: Richardson extrapolation
43 lecture 26: Ricardson extrapolation 35 Ricardson extrapolation, Romberg integration Trougout numerical analysis, one encounters procedures tat apply some simple approximation (eg, linear interpolation)
More informationWYSE Academic Challenge 2004 Sectional Mathematics Solution Set
WYSE Academic Callenge 00 Sectional Matematics Solution Set. Answer: B. Since te equation can be written in te form x + y, we ave a major 5 semi-axis of lengt 5 and minor semi-axis of lengt. Tis means
More informationVolume 29, Issue 3. Existence of competitive equilibrium in economies with multi-member households
Volume 29, Issue 3 Existence of competitive equilibrium in economies wit multi-member ouseolds Noriisa Sato Graduate Scool of Economics, Waseda University Abstract Tis paper focuses on te existence of
More informationAMS 147 Computational Methods and Applications Lecture 09 Copyright by Hongyun Wang, UCSC. Exact value. Effect of round-off error.
Lecture 09 Copyrigt by Hongyun Wang, UCSC Recap: Te total error in numerical differentiation fl( f ( x + fl( f ( x E T ( = f ( x Numerical result from a computer Exact value = e + f x+ Discretization error
More information. If lim. x 2 x 1. f(x+h) f(x)
Review of Differential Calculus Wen te value of one variable y is uniquely determined by te value of anoter variable x, ten te relationsip between x and y is described by a function f tat assigns a value
More informationOrder of Accuracy. ũ h u Ch p, (1)
Order of Accuracy 1 Terminology We consider a numerical approximation of an exact value u. Te approximation depends on a small parameter, wic can be for instance te grid size or time step in a numerical
More information5 Ordinary Differential Equations: Finite Difference Methods for Boundary Problems
5 Ordinary Differential Equations: Finite Difference Metods for Boundary Problems Read sections 10.1, 10.2, 10.4 Review questions 10.1 10.4, 10.8 10.9, 10.13 5.1 Introduction In te previous capters we
More informationA = h w (1) Error Analysis Physics 141
Introduction In all brances of pysical science and engineering one deals constantly wit numbers wic results more or less directly from experimental observations. Experimental observations always ave inaccuracies.
More informationKernel Density Based Linear Regression Estimate
Kernel Density Based Linear Regression Estimate Weixin Yao and Zibiao Zao Abstract For linear regression models wit non-normally distributed errors, te least squares estimate (LSE will lose some efficiency
More information64 IX. The Exceptional Lie Algebras
64 IX. Te Exceptional Lie Algebras IX. Te Exceptional Lie Algebras We ave displayed te four series of classical Lie algebras and teir Dynkin diagrams. How many more simple Lie algebras are tere? Surprisingly,
More informationNew Distribution Theory for the Estimation of Structural Break Point in Mean
New Distribution Teory for te Estimation of Structural Break Point in Mean Liang Jiang Singapore Management University Xiaou Wang Te Cinese University of Hong Kong Jun Yu Singapore Management University
More informationConsider a function f we ll specify which assumptions we need to make about it in a minute. Let us reformulate the integral. 1 f(x) dx.
Capter 2 Integrals as sums and derivatives as differences We now switc to te simplest metods for integrating or differentiating a function from its function samples. A careful study of Taylor expansions
More informationSection 3.1: Derivatives of Polynomials and Exponential Functions
Section 3.1: Derivatives of Polynomials and Exponential Functions In previous sections we developed te concept of te derivative and derivative function. Te only issue wit our definition owever is tat it
More informationExercises for numerical differentiation. Øyvind Ryan
Exercises for numerical differentiation Øyvind Ryan February 25, 2013 1. Mark eac of te following statements as true or false. a. Wen we use te approximation f (a) (f (a +) f (a))/ on a computer, we can
More informationThe Laplace equation, cylindrically or spherically symmetric case
Numerisce Metoden II, 7 4, und Übungen, 7 5 Course Notes, Summer Term 7 Some material and exercises Te Laplace equation, cylindrically or sperically symmetric case Electric and gravitational potential,
More informationSECTION 1.10: DIFFERENCE QUOTIENTS LEARNING OBJECTIVES
(Section.0: Difference Quotients).0. SECTION.0: DIFFERENCE QUOTIENTS LEARNING OBJECTIVES Define average rate of cange (and average velocity) algebraically and grapically. Be able to identify, construct,
More informationHandling Missing Data on Asymmetric Distribution
International Matematical Forum, Vol. 8, 03, no. 4, 53-65 Handling Missing Data on Asymmetric Distribution Amad M. H. Al-Kazale Department of Matematics, Faculty of Science Al-albayt University, Al-Mafraq-Jordan
More informationDELFT UNIVERSITY OF TECHNOLOGY Faculty of Electrical Engineering, Mathematics and Computer Science
DELFT UNIVERSITY OF TECHNOLOGY Faculty of Electrical Engineering, Matematics and Computer Science. ANSWERS OF THE TEST NUMERICAL METHODS FOR DIFFERENTIAL EQUATIONS (WI3097 TU) Tuesday January 9 008, 9:00-:00
More informationSolve exponential equations in one variable using a variety of strategies. LEARN ABOUT the Math. What is the half-life of radon?
8.5 Solving Exponential Equations GOAL Solve exponential equations in one variable using a variety of strategies. LEARN ABOUT te Mat All radioactive substances decrease in mass over time. Jamie works in
More informationCombining functions: algebraic methods
Combining functions: algebraic metods Functions can be added, subtracted, multiplied, divided, and raised to a power, just like numbers or algebra expressions. If f(x) = x 2 and g(x) = x + 2, clearly f(x)
More informationNumerical Analysis MTH603. dy dt = = (0) , y n+1. We obtain yn. Therefore. and. Copyright Virtual University of Pakistan 1
Numerical Analysis MTH60 PREDICTOR CORRECTOR METHOD Te metods presented so far are called single-step metods, were we ave seen tat te computation of y at t n+ tat is y n+ requires te knowledge of y n only.
More informationPolynomial Interpolation
Capter 4 Polynomial Interpolation In tis capter, we consider te important problem of approximatinga function fx, wose values at a set of distinct points x, x, x,, x n are known, by a polynomial P x suc
More informationSection 3: The Derivative Definition of the Derivative
Capter 2 Te Derivative Business Calculus 85 Section 3: Te Derivative Definition of te Derivative Returning to te tangent slope problem from te first section, let's look at te problem of finding te slope
More informationThe derivative function
Roberto s Notes on Differential Calculus Capter : Definition of derivative Section Te derivative function Wat you need to know already: f is at a point on its grap and ow to compute it. Wat te derivative
More informationLecture XVII. Abstract We introduce the concept of directional derivative of a scalar function and discuss its relation with the gradient operator.
Lecture XVII Abstract We introduce te concept of directional derivative of a scalar function and discuss its relation wit te gradient operator. Directional derivative and gradient Te directional derivative
More information2.8 The Derivative as a Function
.8 Te Derivative as a Function Typically, we can find te derivative of a function f at many points of its domain: Definition. Suppose tat f is a function wic is differentiable at every point of an open
More informationSection 15.6 Directional Derivatives and the Gradient Vector
Section 15.6 Directional Derivatives and te Gradient Vector Finding rates of cange in different directions Recall tat wen we first started considering derivatives of functions of more tan one variable,
More informationDifferentiation in higher dimensions
Capter 2 Differentiation in iger dimensions 2.1 Te Total Derivative Recall tat if f : R R is a 1-variable function, and a R, we say tat f is differentiable at x = a if and only if te ratio f(a+) f(a) tends
More informationMathematics 5 Worksheet 11 Geometry, Tangency, and the Derivative
Matematics 5 Workseet 11 Geometry, Tangency, and te Derivative Problem 1. Find te equation of a line wit slope m tat intersects te point (3, 9). Solution. Te equation for a line passing troug a point (x
More informationMATH745 Fall MATH745 Fall
MATH745 Fall 5 MATH745 Fall 5 INTRODUCTION WELCOME TO MATH 745 TOPICS IN NUMERICAL ANALYSIS Instructor: Dr Bartosz Protas Department of Matematics & Statistics Email: bprotas@mcmasterca Office HH 36, Ext
More informationHOMEWORK HELP 2 FOR MATH 151
HOMEWORK HELP 2 FOR MATH 151 Here we go; te second round of omework elp. If tere are oters you would like to see, let me know! 2.4, 43 and 44 At wat points are te functions f(x) and g(x) = xf(x)continuous,
More informationChapter 1. Density Estimation
Capter 1 Density Estimation Let X 1, X,..., X n be observations from a density f X x. Te aim is to use only tis data to obtain an estimate ˆf X x of f X x. Properties of f f X x x, Parametric metods f
More informationThe Priestley-Chao Estimator
Te Priestley-Cao Estimator In tis section we will consider te Pristley-Cao estimator of te unknown regression function. It is assumed tat we ave a sample of observations (Y i, x i ), i = 1,..., n wic are
More informationTaylor Series and the Mean Value Theorem of Derivatives
1 - Taylor Series and te Mean Value Teorem o Derivatives Te numerical solution o engineering and scientiic problems described by matematical models oten requires solving dierential equations. Dierential
More informationlecture 35: Linear Multistep Mehods: Truncation Error
88 lecture 5: Linear Multistep Meods: Truncation Error 5.5 Linear ultistep etods One-step etods construct an approxiate solution x k+ x(t k+ ) using only one previous approxiation, x k. Tis approac enoys
More informationApplications of the van Trees inequality to non-parametric estimation.
Brno-06, Lecture 2, 16.05.06 D/Stat/Brno-06/2.tex www.mast.queensu.ca/ blevit/ Applications of te van Trees inequality to non-parametric estimation. Regular non-parametric problems. As an example of suc
More informationMath 102 TEST CHAPTERS 3 & 4 Solutions & Comments Fall 2006
Mat 102 TEST CHAPTERS 3 & 4 Solutions & Comments Fall 2006 f(x+) f(x) 10 1. For f(x) = x 2 + 2x 5, find ))))))))) and simplify completely. NOTE: **f(x+) is NOT f(x)+! f(x+) f(x) (x+) 2 + 2(x+) 5 ( x 2
More information4. The slope of the line 2x 7y = 8 is (a) 2/7 (b) 7/2 (c) 2 (d) 2/7 (e) None of these.
Mat 11. Test Form N Fall 016 Name. Instructions. Te first eleven problems are wort points eac. Te last six problems are wort 5 points eac. For te last six problems, you must use relevant metods of algebra
More informationBootstrap confidence intervals in nonparametric regression without an additive model
Bootstrap confidence intervals in nonparametric regression witout an additive model Dimitris N. Politis Abstract Te problem of confidence interval construction in nonparametric regression via te bootstrap
More informationIntroduction to Derivatives
Introduction to Derivatives 5-Minute Review: Instantaneous Rates and Tangent Slope Recall te analogy tat we developed earlier First we saw tat te secant slope of te line troug te two points (a, f (a))
More information3.1 Extreme Values of a Function
.1 Etreme Values of a Function Section.1 Notes Page 1 One application of te derivative is finding minimum and maimum values off a grap. In precalculus we were only able to do tis wit quadratics by find
More information1. State whether the function is an exponential growth or exponential decay, and describe its end behaviour using limits.
Questions 1. State weter te function is an exponential growt or exponential decay, and describe its end beaviour using its. (a) f(x) = 3 2x (b) f(x) = 0.5 x (c) f(x) = e (d) f(x) = ( ) x 1 4 2. Matc te
More informationHomework 1 Due: Wednesday, September 28, 2016
0-704 Information Processing and Learning Fall 06 Homework Due: Wednesday, September 8, 06 Notes: For positive integers k, [k] := {,..., k} denotes te set of te first k positive integers. Wen p and Y q
More informationPolynomial Interpolation
Capter 4 Polynomial Interpolation In tis capter, we consider te important problem of approximating a function f(x, wose values at a set of distinct points x, x, x 2,,x n are known, by a polynomial P (x
More informationDigital Filter Structures
Digital Filter Structures Te convolution sum description of an LTI discrete-time system can, in principle, be used to implement te system For an IIR finite-dimensional system tis approac is not practical
More informationTHE STURM-LIOUVILLE-TRANSFORMATION FOR THE SOLUTION OF VECTOR PARTIAL DIFFERENTIAL EQUATIONS. L. Trautmann, R. Rabenstein
Worksop on Transforms and Filter Banks (WTFB),Brandenburg, Germany, Marc 999 THE STURM-LIOUVILLE-TRANSFORMATION FOR THE SOLUTION OF VECTOR PARTIAL DIFFERENTIAL EQUATIONS L. Trautmann, R. Rabenstein Lerstul
More informationHow to Find the Derivative of a Function: Calculus 1
Introduction How to Find te Derivative of a Function: Calculus 1 Calculus is not an easy matematics course Te fact tat you ave enrolled in suc a difficult subject indicates tat you are interested in te
More information, meant to remind us of the definition of f (x) as the limit of difference quotients: = lim
Mat 132 Differentiation Formulas Stewart 2.3 So far, we ave seen ow various real-world problems rate of cange and geometric problems tangent lines lead to derivatives. In tis section, we will see ow to
More informationExam 1 Review Solutions
Exam Review Solutions Please also review te old quizzes, and be sure tat you understand te omework problems. General notes: () Always give an algebraic reason for your answer (graps are not sufficient),
More informationEFFICIENCY OF MODEL-ASSISTED REGRESSION ESTIMATORS IN SAMPLE SURVEYS
Statistica Sinica 24 2014, 395-414 doi:ttp://dx.doi.org/10.5705/ss.2012.064 EFFICIENCY OF MODEL-ASSISTED REGRESSION ESTIMATORS IN SAMPLE SURVEYS Jun Sao 1,2 and Seng Wang 3 1 East Cina Normal University,
More information3.4 Worksheet: Proof of the Chain Rule NAME
Mat 1170 3.4 Workseet: Proof of te Cain Rule NAME Te Cain Rule So far we are able to differentiate all types of functions. For example: polynomials, rational, root, and trigonometric functions. We are
More information7 Semiparametric Methods and Partially Linear Regression
7 Semiparametric Metods and Partially Linear Regression 7. Overview A model is called semiparametric if it is described by and were is nite-dimensional (e.g. parametric) and is in nite-dimensional (nonparametric).
More information1 The concept of limits (p.217 p.229, p.242 p.249, p.255 p.256) 1.1 Limits Consider the function determined by the formula 3. x since at this point
MA00 Capter 6 Calculus and Basic Linear Algebra I Limits, Continuity and Differentiability Te concept of its (p.7 p.9, p.4 p.49, p.55 p.56). Limits Consider te function determined by te formula f Note
More informationSTAT Homework X - Solutions
STAT-36700 Homework X - Solutions Fall 201 November 12, 201 Tis contains solutions for Homework 4. Please note tat we ave included several additional comments and approaces to te problems to give you better
More information158 Calculus and Structures
58 Calculus and Structures CHAPTER PROPERTIES OF DERIVATIVES AND DIFFERENTIATION BY THE EASY WAY. Calculus and Structures 59 Copyrigt Capter PROPERTIES OF DERIVATIVES. INTRODUCTION In te last capter you
More information2.1 THE DEFINITION OF DERIVATIVE
2.1 Te Derivative Contemporary Calculus 2.1 THE DEFINITION OF DERIVATIVE 1 Te grapical idea of a slope of a tangent line is very useful, but for some uses we need a more algebraic definition of te derivative
More informationPolynomials 3: Powers of x 0 + h
near small binomial Capter 17 Polynomials 3: Powers of + Wile it is easy to compute wit powers of a counting-numerator, it is a lot more difficult to compute wit powers of a decimal-numerator. EXAMPLE
More informationSECTION 3.2: DERIVATIVE FUNCTIONS and DIFFERENTIABILITY
(Section 3.2: Derivative Functions and Differentiability) 3.2.1 SECTION 3.2: DERIVATIVE FUNCTIONS and DIFFERENTIABILITY LEARNING OBJECTIVES Know, understand, and apply te Limit Definition of te Derivative
More informationLast lecture (#4): J vortex. J tr
Last lecture (#4): We completed te discussion of te B-T pase diagram of type- and type- superconductors. n contrast to type-, te type- state as finite resistance unless vortices are pinned by defects.
More informationName: Answer Key No calculators. Show your work! 1. (21 points) All answers should either be,, a (finite) real number, or DNE ( does not exist ).
Mat - Final Exam August 3 rd, Name: Answer Key No calculators. Sow your work!. points) All answers sould eiter be,, a finite) real number, or DNE does not exist ). a) Use te grap of te function to evaluate
More informationIEOR 165 Lecture 10 Distribution Estimation
IEOR 165 Lecture 10 Distribution Estimation 1 Motivating Problem Consider a situation were we ave iid data x i from some unknown distribution. One problem of interest is estimating te distribution tat
More informationArtificial Neural Network Model Based Estimation of Finite Population Total
International Journal of Science and Researc (IJSR), India Online ISSN: 2319-7064 Artificial Neural Network Model Based Estimation of Finite Population Total Robert Kasisi 1, Romanus O. Odiambo 2, Antony
More informationBootstrap prediction intervals for Markov processes
arxiv: arxiv:0000.0000 Bootstrap prediction intervals for Markov processes Li Pan and Dimitris N. Politis Li Pan Department of Matematics University of California San Diego La Jolla, CA 92093-0112, USA
More informationQuasiperiodic phenomena in the Van der Pol - Mathieu equation
Quasiperiodic penomena in te Van der Pol - Matieu equation F. Veerman and F. Verulst Department of Matematics, Utrect University P.O. Box 80.010, 3508 TA Utrect Te Neterlands April 8, 009 Abstract Te Van
More informationLogarithmic functions
Roberto s Notes on Differential Calculus Capter 5: Derivatives of transcendental functions Section Derivatives of Logaritmic functions Wat ou need to know alread: Definition of derivative and all basic
More information1 Calculus. 1.1 Gradients and the Derivative. Q f(x+h) f(x)
Calculus. Gradients and te Derivative Q f(x+) δy P T δx R f(x) 0 x x+ Let P (x, f(x)) and Q(x+, f(x+)) denote two points on te curve of te function y = f(x) and let R denote te point of intersection of
More informationTeaching Differentiation: A Rare Case for the Problem of the Slope of the Tangent Line
Teacing Differentiation: A Rare Case for te Problem of te Slope of te Tangent Line arxiv:1805.00343v1 [mat.ho] 29 Apr 2018 Roman Kvasov Department of Matematics University of Puerto Rico at Aguadilla Aguadilla,
More informationTe comparison of dierent models M i is based on teir relative probabilities, wic can be expressed, again using Bayes' teorem, in terms of prior probab
To appear in: Advances in Neural Information Processing Systems 9, eds. M. C. Mozer, M. I. Jordan and T. Petsce. MIT Press, 997 Bayesian Model Comparison by Monte Carlo Caining David Barber D.Barber@aston.ac.uk
More informationThe Derivative as a Function
Section 2.2 Te Derivative as a Function 200 Kiryl Tsiscanka Te Derivative as a Function DEFINITION: Te derivative of a function f at a number a, denoted by f (a), is if tis limit exists. f (a) f(a + )
More informationFinding and Using Derivative The shortcuts
Calculus 1 Lia Vas Finding and Using Derivative Te sortcuts We ave seen tat te formula f f(x+) f(x) (x) = lim 0 is manageable for relatively simple functions like a linear or quadratic. For more complex
More informationMore on generalized inverses of partitioned matrices with Banachiewicz-Schur forms
More on generalized inverses of partitioned matrices wit anaciewicz-scur forms Yongge Tian a,, Yosio Takane b a Cina Economics and Management cademy, Central University of Finance and Economics, eijing,
More informationLIMITATIONS OF EULER S METHOD FOR NUMERICAL INTEGRATION
LIMITATIONS OF EULER S METHOD FOR NUMERICAL INTEGRATION LAURA EVANS.. Introduction Not all differential equations can be explicitly solved for y. Tis can be problematic if we need to know te value of y
More informationCopyright c 2008 Kevin Long
Lecture 4 Numerical solution of initial value problems Te metods you ve learned so far ave obtained closed-form solutions to initial value problems. A closedform solution is an explicit algebriac formula
More informationch (for some fixed positive number c) reaching c
GSTF Journal of Matematics Statistics and Operations Researc (JMSOR) Vol. No. September 05 DOI 0.60/s4086-05-000-z Nonlinear Piecewise-defined Difference Equations wit Reciprocal and Cubic Terms Ramadan
More informationDerivation Of The Schwarzschild Radius Without General Relativity
Derivation Of Te Scwarzscild Radius Witout General Relativity In tis paper I present an alternative metod of deriving te Scwarzscild radius of a black ole. Te metod uses tree of te Planck units formulas:
More information= 0 and states ''hence there is a stationary point'' All aspects of the proof dx must be correct (c)
Paper 1: Pure Matematics 1 Mark Sceme 1(a) (i) (ii) d d y 3 1x 4x x M1 A1 d y dx 1.1b 1.1b 36x 48x A1ft 1.1b Substitutes x = into teir dx (3) 3 1 4 Sows d y 0 and states ''ence tere is a stationary point''
More information3. THE EXCHANGE ECONOMY
Essential Microeconomics -1-3. THE EXCHNGE ECONOMY Pareto efficient allocations 2 Edgewort box analysis 5 Market clearing prices 13 Walrasian Equilibrium 16 Equilibrium and Efficiency 22 First welfare
More informationFast Exact Univariate Kernel Density Estimation
Fast Exact Univariate Kernel Density Estimation David P. Hofmeyr Department of Statistics and Actuarial Science, Stellenbosc University arxiv:1806.00690v2 [stat.co] 12 Jul 2018 July 13, 2018 Abstract Tis
More informationResearch Article New Results on Multiple Solutions for Nth-Order Fuzzy Differential Equations under Generalized Differentiability
Hindawi Publising Corporation Boundary Value Problems Volume 009, Article ID 395714, 13 pages doi:10.1155/009/395714 Researc Article New Results on Multiple Solutions for Nt-Order Fuzzy Differential Equations
More informationInvestigation of Tangent Polynomials with a Computer Algebra System The AMATYC Review, Vol. 14, No. 1, Fall 1992, pp
Investigation of Tangent Polynomials wit a Computer Algebra System Te AMATYC Review, Vol. 14, No. 1, Fall 199, pp. -7. Jon H. Matews California State University Fullerton Fullerton, CA 9834 By Russell
More informationPhysically Based Modeling: Principles and Practice Implicit Methods for Differential Equations
Pysically Based Modeling: Principles and Practice Implicit Metods for Differential Equations David Baraff Robotics Institute Carnegie Mellon University Please note: Tis document is 997 by David Baraff
More informationDefinition of the Derivative
Te Limit Definition of te Derivative Tis Handout will: Define te limit grapically and algebraically Discuss, in detail, specific features of te definition of te derivative Provide a general strategy of
More informationHigher Derivatives. Differentiable Functions
Calculus 1 Lia Vas Higer Derivatives. Differentiable Functions Te second derivative. Te derivative itself can be considered as a function. Te instantaneous rate of cange of tis function is te second derivative.
More informationDepartment of Statistics & Operations Research, Aligarh Muslim University, Aligarh, India
Open Journal of Optimization, 04, 3, 68-78 Publised Online December 04 in SciRes. ttp://www.scirp.org/ournal/oop ttp://dx.doi.org/0.436/oop.04.34007 Compromise Allocation for Combined Ratio Estimates of
More informationSection 2: The Derivative Definition of the Derivative
Capter 2 Te Derivative Applied Calculus 80 Section 2: Te Derivative Definition of te Derivative Suppose we drop a tomato from te top of a 00 foot building and time its fall. Time (sec) Heigt (ft) 0.0 00
More informationThe Verlet Algorithm for Molecular Dynamics Simulations
Cemistry 380.37 Fall 2015 Dr. Jean M. Standard November 9, 2015 Te Verlet Algoritm for Molecular Dynamics Simulations Equations of motion For a many-body system consisting of N particles, Newton's classical
More informationFunction Composition and Chain Rules
Function Composition and s James K. Peterson Department of Biological Sciences and Department of Matematical Sciences Clemson University Marc 8, 2017 Outline 1 Function Composition and Continuity 2 Function
More informationERROR BOUNDS FOR THE METHODS OF GLIMM, GODUNOV AND LEVEQUE BRADLEY J. LUCIER*
EO BOUNDS FO THE METHODS OF GLIMM, GODUNOV AND LEVEQUE BADLEY J. LUCIE* Abstract. Te expected error in L ) attimet for Glimm s sceme wen applied to a scalar conservation law is bounded by + 2 ) ) /2 T
More information6. Non-uniform bending
. Non-uniform bending Introduction Definition A non-uniform bending is te case were te cross-section is not only bent but also seared. It is known from te statics tat in suc a case, te bending moment in
More informationChapter 5 FINITE DIFFERENCE METHOD (FDM)
MEE7 Computer Modeling Tecniques in Engineering Capter 5 FINITE DIFFERENCE METHOD (FDM) 5. Introduction to FDM Te finite difference tecniques are based upon approximations wic permit replacing differential
More informationParameter Fitted Scheme for Singularly Perturbed Delay Differential Equations
International Journal of Applied Science and Engineering 2013. 11, 4: 361-373 Parameter Fitted Sceme for Singularly Perturbed Delay Differential Equations Awoke Andargiea* and Y. N. Reddyb a b Department
More informationBayesian ML Sequence Detection for ISI Channels
Bayesian ML Sequence Detection for ISI Cannels Jill K. Nelson Department of Electrical and Computer Engineering George Mason University Fairfax, VA 030 Email: jnelson@gmu.edu Andrew C. Singer Department
More informationPolynomial Functions. Linear Functions. Precalculus: Linear and Quadratic Functions
Concepts: definition of polynomial functions, linear functions tree representations), transformation of y = x to get y = mx + b, quadratic functions axis of symmetry, vertex, x-intercepts), transformations
More informationTest 2 Review. 1. Find the determinant of the matrix below using (a) cofactor expansion and (b) row reduction. A = 3 2 =
Test Review Find te determinant of te matrix below using (a cofactor expansion and (b row reduction Answer: (a det + = (b Observe R R R R R R R R R Ten det B = (((det Hence det Use Cramer s rule to solve:
More informationMath 102: A Log-jam. f(x+h) f(x) h. = 10 x ( 10 h 1. = 10x+h 10 x h. = 10x 10 h 10 x h. 2. The hyperbolic cosine function is defined by
Mat 102: A Log-jam 1. If f(x) = 10 x, sow tat f(x+) f(x) ( 10 = 10 x ) 1 f(x+) f(x) = 10x+ 10 x = 10x 10 10 x = 10 x ( 10 1 ) 2. Te yperbolic cosine function is defined by cos(x) = ex +e x 2 Te yperbolic
More information